[ 465.000634] env[62183]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62183) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.001070] env[62183]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62183) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.001070] env[62183]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62183) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.001432] env[62183]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 465.093922] env[62183]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62183) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 465.104745] env[62183]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.011s {{(pid=62183) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 465.704811] env[62183]: INFO nova.virt.driver [None req-2e2107ae-692d-460f-9369-96c5cb9258e2 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 465.774213] env[62183]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 465.774416] env[62183]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 465.774491] env[62183]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62183) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 468.926061] env[62183]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-b146d6b2-c0c1-45bb-8448-2e58252cafb3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.942589] env[62183]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62183) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 468.942801] env[62183]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-f2d2a960-1551-443f-8ad0-fca820dc9eb7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.974355] env[62183]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 92b9f. [ 468.974549] env[62183]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.200s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 468.975033] env[62183]: INFO nova.virt.vmwareapi.driver [None req-2e2107ae-692d-460f-9369-96c5cb9258e2 None None] VMware vCenter version: 7.0.3 [ 468.978413] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3faa8d09-418b-42c2-b286-df9b848e82de {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.996441] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ef954f-ef1a-492e-9c63-352ea502296f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.002335] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a39aa6a-d594-43e2-b0a8-768a402e2a15 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.008913] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65097337-a227-4da1-8116-ffd4caa6b26c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.021802] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e965ff5-53f0-4355-97cd-4e9951563fc7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.027484] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db495e85-81f2-4412-9183-52e8642bf43f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.057494] env[62183]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-2ed0e975-3e06-4b33-aa36-c4addb0c7306 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.062524] env[62183]: DEBUG nova.virt.vmwareapi.driver [None req-2e2107ae-692d-460f-9369-96c5cb9258e2 None None] Extension org.openstack.compute already exists. {{(pid=62183) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 469.065253] env[62183]: INFO nova.compute.provider_config [None req-2e2107ae-692d-460f-9369-96c5cb9258e2 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 469.571023] env[62183]: DEBUG nova.context [None req-2e2107ae-692d-460f-9369-96c5cb9258e2 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),0024293e-e520-41b5-a74a-a3e8b2c17058(cell1) {{(pid=62183) load_cells /opt/stack/nova/nova/context.py:464}} [ 469.571320] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 469.571706] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 469.572578] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 469.573217] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Acquiring lock "0024293e-e520-41b5-a74a-a3e8b2c17058" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 469.573561] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Lock "0024293e-e520-41b5-a74a-a3e8b2c17058" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 469.574705] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Lock "0024293e-e520-41b5-a74a-a3e8b2c17058" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 469.595282] env[62183]: INFO dbcounter [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Registered counter for database nova_cell0 [ 469.604284] env[62183]: INFO dbcounter [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Registered counter for database nova_cell1 [ 469.607803] env[62183]: DEBUG oslo_db.sqlalchemy.engines [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62183) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 469.611022] env[62183]: DEBUG oslo_db.sqlalchemy.engines [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62183) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 469.613373] env[62183]: ERROR nova.db.main.api [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 469.613373] env[62183]: result = function(*args, **kwargs) [ 469.613373] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 469.613373] env[62183]: return func(*args, **kwargs) [ 469.613373] env[62183]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 469.613373] env[62183]: result = fn(*args, **kwargs) [ 469.613373] env[62183]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 469.613373] env[62183]: return f(*args, **kwargs) [ 469.613373] env[62183]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 469.613373] env[62183]: return db.service_get_minimum_version(context, binaries) [ 469.613373] env[62183]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 469.613373] env[62183]: _check_db_access() [ 469.613373] env[62183]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 469.613373] env[62183]: stacktrace = ''.join(traceback.format_stack()) [ 469.613373] env[62183]: [ 469.614782] env[62183]: ERROR nova.db.main.api [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 469.614782] env[62183]: result = function(*args, **kwargs) [ 469.614782] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 469.614782] env[62183]: return func(*args, **kwargs) [ 469.614782] env[62183]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 469.614782] env[62183]: result = fn(*args, **kwargs) [ 469.614782] env[62183]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 469.614782] env[62183]: return f(*args, **kwargs) [ 469.614782] env[62183]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 469.614782] env[62183]: return db.service_get_minimum_version(context, binaries) [ 469.614782] env[62183]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 469.614782] env[62183]: _check_db_access() [ 469.614782] env[62183]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 469.614782] env[62183]: stacktrace = ''.join(traceback.format_stack()) [ 469.614782] env[62183]: [ 469.615760] env[62183]: WARNING nova.objects.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 469.619023] env[62183]: WARNING nova.objects.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Failed to get minimum service version for cell 0024293e-e520-41b5-a74a-a3e8b2c17058 [ 469.619023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Acquiring lock "singleton_lock" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 469.619023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Acquired lock "singleton_lock" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 469.619023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Releasing lock "singleton_lock" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 469.619023] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Full set of CONF: {{(pid=62183) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 469.619023] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ******************************************************************************** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 469.619023] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Configuration options gathered from: {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 469.619267] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 469.619267] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 469.619267] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ================================================================================ {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 469.619267] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] allow_resize_to_same_host = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.619267] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] arq_binding_timeout = 300 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.619267] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] backdoor_port = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.619441] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] backdoor_socket = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.619441] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] block_device_allocate_retries = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.619441] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] block_device_allocate_retries_interval = 3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.619609] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cert = self.pem {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.619906] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.620213] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute_monitors = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.620518] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] config_dir = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.620815] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] config_drive_format = iso9660 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.621076] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.621354] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] config_source = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.621645] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] console_host = devstack {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.621931] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] control_exchange = nova {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.622229] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cpu_allocation_ratio = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.622501] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] daemon = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.622796] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] debug = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.623107] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] default_access_ip_network_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.623486] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] default_availability_zone = nova {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.623863] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] default_ephemeral_format = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.624281] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] default_green_pool_size = 1000 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.624650] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.624981] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] default_schedule_zone = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.625307] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] disk_allocation_ratio = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.625621] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] enable_new_services = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.625924] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] enabled_apis = ['osapi_compute'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.626218] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] enabled_ssl_apis = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.626495] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] flat_injected = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.626768] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] force_config_drive = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.627063] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] force_raw_images = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.627361] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] graceful_shutdown_timeout = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.627634] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] heal_instance_info_cache_interval = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632017] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] host = cpu-1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632017] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632017] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632017] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632017] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632017] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] instance_build_timeout = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632017] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] instance_delete_interval = 300 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632271] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] instance_format = [instance: %(uuid)s] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632271] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] instance_name_template = instance-%08x {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632271] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] instance_usage_audit = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632271] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] instance_usage_audit_period = month {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632271] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632271] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632271] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] internal_service_availability_zone = internal {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632543] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] key = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632543] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] live_migration_retry_count = 30 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632543] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] log_color = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632543] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] log_config_append = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632543] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632543] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] log_dir = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632543] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] log_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632742] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] log_options = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.632786] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] log_rotate_interval = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.633083] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] log_rotate_interval_type = days {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.633393] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] log_rotation_type = none {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.633644] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.633885] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.634329] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.634645] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.634899] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.635195] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] long_rpc_timeout = 1800 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.635471] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] max_concurrent_builds = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.635744] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] max_concurrent_live_migrations = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.636026] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] max_concurrent_snapshots = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.636296] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] max_local_block_devices = 3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.636574] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] max_logfile_count = 30 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.636845] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] max_logfile_size_mb = 200 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.637129] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] maximum_instance_delete_attempts = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.637401] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] metadata_listen = 0.0.0.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.637678] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] metadata_listen_port = 8775 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.637970] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] metadata_workers = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.638272] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] migrate_max_retries = -1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.638566] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] mkisofs_cmd = genisoimage {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.638910] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.639225] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] my_ip = 10.180.1.21 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.639537] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] network_allocate_retries = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.639884] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.640220] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.641023] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] osapi_compute_listen_port = 8774 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.641023] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] osapi_compute_unique_server_name_scope = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.641023] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] osapi_compute_workers = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.641358] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] password_length = 12 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.641666] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] periodic_enable = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.641974] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] periodic_fuzzy_delay = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.642307] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] pointer_model = usbtablet {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.642634] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] preallocate_images = none {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.644162] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] publish_errors = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.644162] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] pybasedir = /opt/stack/nova {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.644162] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ram_allocation_ratio = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.644162] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] rate_limit_burst = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.644162] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] rate_limit_except_level = CRITICAL {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.644162] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] rate_limit_interval = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.644162] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] reboot_timeout = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.644610] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] reclaim_instance_interval = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.644913] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] record = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.645255] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] reimage_timeout_per_gb = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.645576] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] report_interval = 120 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.645884] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] rescue_timeout = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.646218] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] reserved_host_cpus = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.646531] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] reserved_host_disk_mb = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.646860] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] reserved_host_memory_mb = 512 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.647185] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] reserved_huge_pages = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.647507] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] resize_confirm_window = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.647835] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] resize_fs_using_block_device = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.648171] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] resume_guests_state_on_host_boot = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.648492] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.649641] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] rpc_response_timeout = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.649641] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] run_external_periodic_tasks = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.649641] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] running_deleted_instance_action = reap {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.649641] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.649641] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] running_deleted_instance_timeout = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.649869] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler_instance_sync_interval = 120 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.650047] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_down_time = 720 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.650290] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] servicegroup_driver = db {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.650512] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] shell_completion = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.650734] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] shelved_offload_time = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.650984] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] shelved_poll_interval = 3600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.651274] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] shutdown_timeout = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.651512] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] source_is_ipv6 = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.651752] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ssl_only = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.652077] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.652319] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] sync_power_state_interval = 600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.652545] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] sync_power_state_pool_size = 1000 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.652775] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] syslog_log_facility = LOG_USER {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.652994] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] tempdir = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.653236] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] timeout_nbd = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.653467] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] transport_url = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.653689] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] update_resources_interval = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.653911] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] use_cow_images = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.654168] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] use_eventlog = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.654537] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] use_journal = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.654790] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] use_json = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.655033] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] use_rootwrap_daemon = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.655266] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] use_stderr = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.655487] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] use_syslog = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.655704] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vcpu_pin_set = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.655933] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plugging_is_fatal = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.656177] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plugging_timeout = 300 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.656410] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] virt_mkfs = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.656634] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] volume_usage_poll_interval = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.656881] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] watch_log_file = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.657128] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] web = /usr/share/spice-html5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.657386] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.657616] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.657856] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658104] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_concurrency.disable_process_locking = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658470] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658716] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658976] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.659244] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.659483] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.659736] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.659974] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.auth_strategy = keystone {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660229] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.compute_link_prefix = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660477] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660727] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.dhcp_domain = novalocal {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660973] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.enable_instance_password = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661232] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.glance_link_prefix = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661459] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661695] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661921] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.instance_list_per_project_cells = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662160] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.list_records_by_skipping_down_cells = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662372] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.local_metadata_per_cell = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662615] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.max_limit = 1000 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662854] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.metadata_cache_expiration = 15 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663125] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.neutron_default_tenant_id = default {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663371] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.response_validation = warn {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663607] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.use_neutron_default_nets = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663850] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664095] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664382] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664706] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664959] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.vendordata_dynamic_targets = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665204] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.vendordata_jsonfile_path = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665408] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665610] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.backend = dogpile.cache.memcached {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665784] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.backend_argument = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665986] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.config_prefix = cache.oslo {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666189] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.dead_timeout = 60.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666362] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.debug_cache_backend = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666530] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.enable_retry_client = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666696] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.enable_socket_keepalive = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666879] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.enabled = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667062] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.enforce_fips_mode = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667239] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.expiration_time = 600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667409] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.hashclient_retry_attempts = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667578] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667744] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_dead_retry = 300 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667907] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_password = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.668227] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.668464] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.668694] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_pool_maxsize = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.668960] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.669232] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_sasl_enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.669483] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.669715] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.669937] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.memcache_username = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670180] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.proxies = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670413] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.redis_db = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670655] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.redis_password = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670886] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671154] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671413] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.redis_server = localhost:6379 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671645] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.redis_socket_timeout = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671867] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.redis_username = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.672130] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.retry_attempts = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.672382] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.retry_delay = 0.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.672613] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.socket_keepalive_count = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.672841] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.socket_keepalive_idle = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673078] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.socket_keepalive_interval = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673307] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.tls_allowed_ciphers = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673527] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.tls_cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673746] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.tls_certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673966] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.tls_enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.674204] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cache.tls_keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.674493] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.674811] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.auth_type = password {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.675087] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.675346] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.675586] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.675816] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.676053] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.cross_az_attach = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.676287] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.debug = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.676508] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.endpoint_template = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.676733] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.http_retries = 3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.676959] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.677195] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.677435] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.os_region_name = RegionOne {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.677661] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.677884] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cinder.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.678147] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.678395] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.cpu_dedicated_set = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.678618] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.cpu_shared_set = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.678869] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.image_type_exclude_list = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.679115] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.679348] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.679572] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.679801] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.680036] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.680307] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.resource_provider_association_refresh = 300 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.680631] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.680897] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.shutdown_retry_interval = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.681182] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.681455] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] conductor.workers = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.681706] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] console.allowed_origins = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.681933] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] console.ssl_ciphers = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.682183] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] console.ssl_minimum_version = default {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.682424] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] consoleauth.enforce_session_timeout = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.682813] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] consoleauth.token_ttl = 600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.682901] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.683087] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.683321] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.683544] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.connect_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.683762] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.connect_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.683983] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.endpoint_override = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.684249] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.684491] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.684832] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.max_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.685090] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.min_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.685322] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.region_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.685543] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.retriable_status_codes = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.685758] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.service_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.685988] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.service_type = accelerator {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.686229] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.686447] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.status_code_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.686662] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.status_code_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.686895] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.687161] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.687406] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] cyborg.version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.687651] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.backend = sqlalchemy {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.687884] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.connection = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.688125] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.connection_debug = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.688364] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.connection_parameters = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.688586] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.connection_recycle_time = 3600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.688832] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.connection_trace = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.689076] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.db_inc_retry_interval = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.689312] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.db_max_retries = 20 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.689557] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.db_max_retry_interval = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.689835] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.db_retry_interval = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.690101] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.max_overflow = 50 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.690337] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.max_pool_size = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.690572] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.max_retries = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.690806] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.691039] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.mysql_wsrep_sync_wait = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.691268] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.pool_timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.691493] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.retry_interval = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.691710] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.slave_connection = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.691931] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.sqlite_synchronous = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.692176] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] database.use_db_reconnect = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.692422] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.backend = sqlalchemy {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.692610] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.connection = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.692783] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.connection_debug = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.693034] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.connection_parameters = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.693252] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.connection_recycle_time = 3600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.693467] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.connection_trace = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.693669] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.db_inc_retry_interval = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.693869] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.db_max_retries = 20 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.694080] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.db_max_retry_interval = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.694286] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.db_retry_interval = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.694486] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.max_overflow = 50 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.694751] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.max_pool_size = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.695045] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.max_retries = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.695313] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.695548] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.695752] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.pool_timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.695956] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.retry_interval = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.696196] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.slave_connection = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.696402] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] api_database.sqlite_synchronous = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.696619] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] devices.enabled_mdev_types = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.696840] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.697061] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.697271] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ephemeral_storage_encryption.enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.697472] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.697680] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.api_servers = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.697861] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.698043] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.698215] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.698403] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.connect_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.698572] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.connect_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.698737] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.debug = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.698937] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.default_trusted_certificate_ids = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.699127] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.enable_certificate_validation = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.699296] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.enable_rbd_download = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.699460] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.endpoint_override = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.699630] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.699797] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.699960] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.max_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.700207] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.min_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.700416] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.num_retries = 3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.700624] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.rbd_ceph_conf = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.700827] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.rbd_connect_timeout = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.701071] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.rbd_pool = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.701315] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.rbd_user = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.701534] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.region_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.701737] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.retriable_status_codes = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.701937] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.service_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.702160] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.service_type = image {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.702366] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.702569] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.status_code_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.702765] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.status_code_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.702962] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.703200] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.703404] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.verify_glance_signatures = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.703599] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] glance.version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.703815] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] guestfs.debug = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.704036] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.704248] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.auth_type = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.704469] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.704673] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.704953] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.705250] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.connect_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.705456] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.connect_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.705653] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.endpoint_override = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.705852] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.706060] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.706263] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.max_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.706468] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.min_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.706665] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.region_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.706879] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.retriable_status_codes = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.707099] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.service_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.707317] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.service_type = shared-file-system {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.707520] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.share_apply_policy_timeout = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.707718] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.707911] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.status_code_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.708119] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.status_code_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.708320] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.708542] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.708741] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] manila.version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.709008] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] mks.enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.709425] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.709660] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] image_cache.manager_interval = 2400 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.709873] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] image_cache.precache_concurrency = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.710095] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] image_cache.remove_unused_base_images = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.710314] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.710521] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.710739] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] image_cache.subdirectory_name = _base {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.710952] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.api_max_retries = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.711196] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.api_retry_interval = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.711413] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.711618] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.auth_type = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.711817] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.712039] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.712252] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.712456] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.conductor_group = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.712652] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.connect_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.712847] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.connect_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.713053] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.endpoint_override = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.713260] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.713454] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.713645] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.max_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.713837] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.min_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.714050] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.peer_list = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.714652] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.region_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.714652] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.retriable_status_codes = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.714652] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.serial_console_state_timeout = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.714783] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.service_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.715074] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.service_type = baremetal {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.715362] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.shard = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.715575] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.715775] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.status_code_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.715970] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.status_code_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.716186] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.716408] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.716609] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ironic.version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.716850] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.717092] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] key_manager.fixed_key = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.717340] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.717547] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.barbican_api_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.717748] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.barbican_endpoint = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.717961] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.barbican_endpoint_type = public {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.718178] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.barbican_region_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.718378] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.718577] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.718782] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.719015] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.719225] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.719432] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.number_of_retries = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.719632] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.retry_delay = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.719865] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.send_service_user_token = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.720115] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.720328] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.720534] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.verify_ssl = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.720757] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican.verify_ssl_path = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.720930] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican_service_user.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.721144] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican_service_user.auth_type = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.721348] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican_service_user.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.721545] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican_service_user.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.721745] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican_service_user.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.721942] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican_service_user.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.722156] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican_service_user.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.724396] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican_service_user.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.724396] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] barbican_service_user.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.724396] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.approle_role_id = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.724396] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.approle_secret_id = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.724396] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.kv_mountpoint = secret {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.724584] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.kv_path = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.724674] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.kv_version = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.724842] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.namespace = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.725080] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.root_token_id = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.725266] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.ssl_ca_crt_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.725443] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.timeout = 60.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.725617] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.use_ssl = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.725795] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.725974] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.726158] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.auth_type = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.726323] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.726486] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.726652] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.726842] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.connect_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.727039] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.connect_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.727212] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.endpoint_override = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.727381] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.727545] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.727705] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.max_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.727867] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.min_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.728040] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.region_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.728210] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.retriable_status_codes = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.728369] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.service_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.728544] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.service_type = identity {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.728708] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.728900] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.status_code_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.729098] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.status_code_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.729268] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.729457] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.729621] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] keystone.version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.729872] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.connection_uri = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.730078] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.cpu_mode = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.730259] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.730433] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.cpu_models = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.730611] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.cpu_power_governor_high = performance {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.730784] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.730952] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.cpu_power_management = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.731142] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.731312] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.device_detach_attempts = 8 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.731492] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.device_detach_timeout = 20 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.731657] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.disk_cachemodes = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.731819] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.disk_prefix = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.731989] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.enabled_perf_events = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.732174] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.file_backed_memory = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.732344] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.gid_maps = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.732508] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.hw_disk_discard = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.732672] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.hw_machine_type = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.732844] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.images_rbd_ceph_conf = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.733047] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.733222] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.733395] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.images_rbd_glance_store_name = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.733566] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.images_rbd_pool = rbd {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.733738] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.images_type = default {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.733902] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.images_volume_group = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.734131] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.inject_key = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.734312] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.inject_partition = -2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.734477] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.inject_password = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.734644] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.iscsi_iface = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.734808] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.iser_use_multipath = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.734976] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.735157] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.735325] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_downtime = 500 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.735488] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.735650] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.735812] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_inbound_addr = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.736014] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.736194] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.736356] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_scheme = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.736528] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_timeout_action = abort {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.736694] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_tunnelled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.736855] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_uri = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.737030] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.live_migration_with_native_tls = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.737253] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.max_queues = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.737432] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.737671] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.737840] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.nfs_mount_options = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.738165] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.738345] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.738515] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.738680] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.738866] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.739055] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.num_pcie_ports = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.739230] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.739401] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.pmem_namespaces = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.739559] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.quobyte_client_cfg = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.739877] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.740076] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.740254] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.740427] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.740593] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rbd_secret_uuid = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.740760] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rbd_user = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.740927] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.741121] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.741287] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rescue_image_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.741452] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rescue_kernel_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.741617] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rescue_ramdisk_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.741779] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.741943] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.rx_queue_size = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.742132] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.smbfs_mount_options = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.742412] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.742586] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.snapshot_compression = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.742755] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.snapshot_image_format = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.742999] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.743187] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.sparse_logical_volumes = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.743355] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.swtpm_enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.743528] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.swtpm_group = tss {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.743701] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.swtpm_user = tss {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.743874] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.sysinfo_serial = unique {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.744051] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.tb_cache_size = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.744218] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.tx_queue_size = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.744386] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.uid_maps = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.744550] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.use_virtio_for_bridges = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.744723] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.virt_type = kvm {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.744906] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.volume_clear = zero {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.745086] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.volume_clear_size = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.745261] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.volume_use_multipath = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.745440] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.vzstorage_cache_path = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.745616] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.745788] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.745958] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.746172] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.746504] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.746709] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.vzstorage_mount_user = stack {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.746885] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.747077] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.747257] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.auth_type = password {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.747422] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.747581] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.747746] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.747906] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.connect_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.748079] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.connect_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.748255] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.default_floating_pool = public {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.748417] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.endpoint_override = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.748580] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.extension_sync_interval = 600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.748741] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.http_retries = 3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.748928] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.749109] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.749274] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.max_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.749506] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.749687] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.min_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.749865] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.ovs_bridge = br-int {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.750045] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.physnets = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.750224] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.region_name = RegionOne {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.750389] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.retriable_status_codes = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.750564] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.service_metadata_proxy = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.750728] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.service_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.750897] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.service_type = network {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.751076] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.751246] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.status_code_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.751405] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.status_code_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.751627] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.751775] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.751940] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] neutron.version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.752135] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] notifications.bdms_in_notifications = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.752317] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] notifications.default_level = INFO {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.752494] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] notifications.notification_format = unversioned {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.752659] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] notifications.notify_on_state_change = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.752837] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.753024] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] pci.alias = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.753200] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] pci.device_spec = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.753366] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] pci.report_in_placement = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.753542] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.753716] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.auth_type = password {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.753884] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.754058] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.754223] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.754387] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.754547] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.connect_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.754706] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.connect_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.754868] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.default_domain_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.755037] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.default_domain_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.755201] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.domain_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.755361] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.domain_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.755520] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.endpoint_override = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.755684] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.755843] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.756008] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.max_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.756178] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.min_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.756350] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.password = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.756509] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.project_domain_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.756676] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.project_domain_name = Default {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.756865] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.project_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.757068] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.project_name = service {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.757248] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.region_name = RegionOne {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.757415] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.retriable_status_codes = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.757581] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.service_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.757753] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.service_type = placement {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.757919] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.758094] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.status_code_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.758260] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.status_code_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.758423] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.system_scope = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.758582] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.758742] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.trust_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.758934] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.user_domain_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.759129] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.user_domain_name = Default {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.759294] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.user_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.759471] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.username = nova {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.759654] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.759835] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] placement.version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.760038] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.cores = 20 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.760212] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.count_usage_from_placement = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.760387] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.760564] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.injected_file_content_bytes = 10240 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.760734] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.injected_file_path_length = 255 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.760902] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.injected_files = 5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.761086] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.instances = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.761261] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.key_pairs = 100 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.761430] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.metadata_items = 128 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.761596] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.ram = 51200 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.761766] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.recheck_quota = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.761929] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.server_group_members = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.762108] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] quota.server_groups = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.762283] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.762446] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.762606] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.image_metadata_prefilter = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.762765] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.762925] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.max_attempts = 3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.763103] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.max_placement_results = 1000 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.763267] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.763428] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.763587] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.763761] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] scheduler.workers = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.763935] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.764130] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.764312] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.764482] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.764650] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.764814] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.764982] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.765187] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.765359] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.host_subset_size = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.765528] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.765692] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.765860] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.766054] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.isolated_hosts = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.766250] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.isolated_images = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.766421] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.766585] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.766750] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.766942] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.pci_in_placement = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.767137] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.767372] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.767460] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.767622] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.767803] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.767967] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.768146] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.track_instance_changes = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.768328] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.768499] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] metrics.required = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.768667] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] metrics.weight_multiplier = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.768850] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.769040] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] metrics.weight_setting = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.769362] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.769538] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] serial_console.enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.769715] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] serial_console.port_range = 10000:20000 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.769911] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.770103] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.770277] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] serial_console.serialproxy_port = 6083 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.770444] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.770617] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.auth_type = password {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.770777] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.770933] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.771111] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.771275] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.771432] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.771603] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.send_service_user_token = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.771830] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.772010] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] service_user.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.772197] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.agent_enabled = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.772363] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.772671] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.772863] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.773050] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.html5proxy_port = 6082 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.773215] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.image_compression = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.773376] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.jpeg_compression = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.773535] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.playback_compression = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.773698] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.require_secure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.773869] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.server_listen = 127.0.0.1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.774052] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.774218] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.streaming_mode = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.774379] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] spice.zlib_compression = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.774541] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] upgrade_levels.baseapi = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.774712] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] upgrade_levels.compute = auto {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.774873] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] upgrade_levels.conductor = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.775047] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] upgrade_levels.scheduler = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.775219] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.775382] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.775542] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vendordata_dynamic_auth.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.775702] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vendordata_dynamic_auth.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.775865] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.776036] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vendordata_dynamic_auth.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.776202] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.776362] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.776521] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vendordata_dynamic_auth.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.776697] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.api_retry_count = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.776876] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.ca_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.777072] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.777247] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.cluster_name = testcl1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.777411] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.connection_pool_size = 10 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.777569] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.console_delay_seconds = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.777736] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.datastore_regex = ^datastore.* {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.777942] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.778137] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.host_password = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.778307] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.host_port = 443 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.778491] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.host_username = administrator@vsphere.local {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.778659] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.insecure = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.778841] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.integration_bridge = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.779034] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.maximum_objects = 100 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.779339] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.pbm_default_policy = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.779396] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.pbm_enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.779598] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.pbm_wsdl_location = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.779774] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.779939] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.serial_port_proxy_uri = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.780112] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.serial_port_service_uri = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.780278] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.task_poll_interval = 0.5 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.780450] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.use_linked_clone = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.780620] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.vnc_keymap = en-us {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.780788] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.vnc_port = 5900 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.780953] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vmware.vnc_port_total = 10000 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.781159] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.auth_schemes = ['none'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.781337] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.781634] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.781823] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.782072] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.novncproxy_port = 6080 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.782234] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.server_listen = 127.0.0.1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.782411] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.782584] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.vencrypt_ca_certs = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.782747] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.vencrypt_client_cert = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.782904] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vnc.vencrypt_client_key = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.783090] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.783262] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.disable_deep_image_inspection = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.783421] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.783581] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.783740] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.783902] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.disable_rootwrap = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.784077] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.enable_numa_live_migration = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.784242] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.784404] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.784565] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.784726] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.libvirt_disable_apic = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.784885] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.785060] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.785226] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.785388] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.785547] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.785706] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.785866] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.786036] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.786201] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.786366] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.786609] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.786808] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.client_socket_timeout = 900 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.786980] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.default_pool_size = 1000 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.787168] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.keep_alive = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.787339] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.max_header_line = 16384 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.787503] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.787672] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.ssl_ca_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.787828] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.ssl_cert_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.787990] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.ssl_key_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.788172] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.tcp_keepidle = 600 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.788360] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.788525] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] zvm.ca_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.788684] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] zvm.cloud_connector_url = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.789009] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.789203] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] zvm.reachable_timeout = 300 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.789391] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.enforce_new_defaults = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.789792] env[62183]: WARNING oslo_config.cfg [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 469.789990] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.enforce_scope = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.790188] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.policy_default_rule = default {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.790373] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.790553] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.policy_file = policy.yaml {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.790728] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.790895] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.791072] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.791238] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.791403] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.791572] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.791751] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.791933] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.connection_string = messaging:// {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.792124] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.enabled = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.792295] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.es_doc_type = notification {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.792462] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.es_scroll_size = 10000 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.792631] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.es_scroll_time = 2m {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.792796] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.filter_error_trace = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.792970] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.hmac_keys = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.793153] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.sentinel_service_name = mymaster {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.793322] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.socket_timeout = 0.1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.793482] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.trace_requests = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.793642] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler.trace_sqlalchemy = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.793828] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler_jaeger.process_tags = {} {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.793991] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler_jaeger.service_name_prefix = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.794167] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] profiler_otlp.service_name_prefix = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.794336] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] remote_debug.host = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.794496] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] remote_debug.port = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.794678] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.794845] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.795019] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.795189] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.795354] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.795513] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.795675] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.795839] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.796019] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.796190] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.796350] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.796521] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.796693] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.796887] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.797089] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.797265] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.797433] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.797610] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.797773] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.797937] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.798119] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.798285] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.798446] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.798614] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.798775] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.798963] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.799144] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.799312] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.799478] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.799646] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.ssl = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.799842] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.800047] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.800223] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.800397] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.800568] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.800733] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.800920] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.801105] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_notifications.retry = -1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.801298] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.801480] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.801648] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.auth_section = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.801810] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.auth_type = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.801968] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.cafile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.802155] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.certfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.802305] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.collect_timing = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.802455] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.connect_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.802612] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.connect_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.802765] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.endpoint_id = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.802922] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.endpoint_override = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.803086] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.insecure = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.803247] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.keyfile = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.803404] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.max_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.803559] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.min_version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.803711] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.region_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.803870] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.retriable_status_codes = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.804034] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.service_name = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.804195] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.service_type = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.804356] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.split_loggers = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.804512] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.status_code_retries = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.804667] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.status_code_retry_delay = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.804827] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.timeout = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.804985] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.valid_interfaces = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.805153] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_limit.version = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.805316] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_reports.file_event_handler = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.805480] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.805640] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] oslo_reports.log_dir = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.805806] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.805969] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.806164] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.806335] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.806499] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.806656] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.806844] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.807034] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_ovs_privileged.group = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.807202] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.807370] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.807535] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.807694] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] vif_plug_ovs_privileged.user = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.807868] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.808060] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.808242] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.808417] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.808588] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.808763] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.808955] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.809141] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.809325] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.809499] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_ovs.isolate_vif = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.809668] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.809857] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.810058] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.810239] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.810406] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] os_vif_ovs.per_port_bridge = False {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.810581] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] privsep_osbrick.capabilities = [21] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.810736] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] privsep_osbrick.group = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.810896] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] privsep_osbrick.helper_command = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.811074] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.811244] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.811406] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] privsep_osbrick.user = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.811580] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.811742] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] nova_sys_admin.group = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.811899] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] nova_sys_admin.helper_command = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.812075] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.812247] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.812396] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] nova_sys_admin.user = None {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.812528] env[62183]: DEBUG oslo_service.service [None req-07b7bf2c-5530-4ba8-9324-783ed63ad871 None None] ******************************************************************************** {{(pid=62183) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 469.813017] env[62183]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 470.316422] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Getting list of instances from cluster (obj){ [ 470.316422] env[62183]: value = "domain-c8" [ 470.316422] env[62183]: _type = "ClusterComputeResource" [ 470.316422] env[62183]: } {{(pid=62183) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 470.317933] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57db560d-521a-42b6-9ef9-fd4d3e483cf0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.326964] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Got total of 0 instances {{(pid=62183) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 470.327505] env[62183]: WARNING nova.virt.vmwareapi.driver [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 470.327972] env[62183]: INFO nova.virt.node [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Generated node identity 09c07e5d-2ed9-41c2-be62-db0f731d0b87 [ 470.328222] env[62183]: INFO nova.virt.node [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Wrote node identity 09c07e5d-2ed9-41c2-be62-db0f731d0b87 to /opt/stack/data/n-cpu-1/compute_id [ 470.830879] env[62183]: WARNING nova.compute.manager [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Compute nodes ['09c07e5d-2ed9-41c2-be62-db0f731d0b87'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 471.836357] env[62183]: INFO nova.compute.manager [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 472.842416] env[62183]: WARNING nova.compute.manager [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 472.842909] env[62183]: DEBUG oslo_concurrency.lockutils [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 472.842909] env[62183]: DEBUG oslo_concurrency.lockutils [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 472.843061] env[62183]: DEBUG oslo_concurrency.lockutils [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 472.843225] env[62183]: DEBUG nova.compute.resource_tracker [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 472.844183] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8f0a97-d92e-443e-a5b1-f055338e68e8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.852689] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec6e9b8-0c29-41c3-b356-17989fb86bf6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.867144] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07448040-1828-44d4-8b10-e0608633a5f6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.874597] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4adbb5e3-a372-41e1-a427-0bd58f6f3781 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.905655] env[62183]: DEBUG nova.compute.resource_tracker [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181374MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 472.905845] env[62183]: DEBUG oslo_concurrency.lockutils [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 472.906023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 473.408677] env[62183]: WARNING nova.compute.resource_tracker [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] No compute node record for cpu-1:09c07e5d-2ed9-41c2-be62-db0f731d0b87: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 09c07e5d-2ed9-41c2-be62-db0f731d0b87 could not be found. [ 473.912857] env[62183]: INFO nova.compute.resource_tracker [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 [ 475.420917] env[62183]: DEBUG nova.compute.resource_tracker [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 475.421376] env[62183]: DEBUG nova.compute.resource_tracker [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 475.575771] env[62183]: INFO nova.scheduler.client.report [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] [req-04269880-a0cd-4814-957e-9a433289c43d] Created resource provider record via placement API for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 475.591701] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd0215f-2ebd-4576-a91a-10a3992c34c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.599430] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc7e2e7-ba69-4c5e-8e67-45fae70483d1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.628327] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14151cb-93a1-45d2-9101-4bb37c1cc043 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.635322] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a35361-9de2-4902-a1d3-7531329e3ecd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.647718] env[62183]: DEBUG nova.compute.provider_tree [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.185094] env[62183]: DEBUG nova.scheduler.client.report [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 476.185406] env[62183]: DEBUG nova.compute.provider_tree [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 0 to 1 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 476.185565] env[62183]: DEBUG nova.compute.provider_tree [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.236717] env[62183]: DEBUG nova.compute.provider_tree [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 1 to 2 during operation: update_traits {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 476.744623] env[62183]: DEBUG nova.compute.resource_tracker [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 476.745067] env[62183]: DEBUG oslo_concurrency.lockutils [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.839s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 476.745067] env[62183]: DEBUG nova.service [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Creating RPC server for service compute {{(pid=62183) start /opt/stack/nova/nova/service.py:186}} [ 476.759998] env[62183]: DEBUG nova.service [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] Join ServiceGroup membership for this service compute {{(pid=62183) start /opt/stack/nova/nova/service.py:203}} [ 476.760211] env[62183]: DEBUG nova.servicegroup.drivers.db [None req-984f86c6-3b1a-43c4-ad37-e190522037bd None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62183) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 516.482876] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Acquiring lock "53dd2793-a2da-45e4-8d25-c9b8875c0402" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.483206] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Lock "53dd2793-a2da-45e4-8d25-c9b8875c0402" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.985780] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.513672] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Acquiring lock "495347c6-1078-4a74-96b7-4ec0099cc352" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.514028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Lock "495347c6-1078-4a74-96b7-4ec0099cc352" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.540614] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.540614] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.541620] env[62183]: INFO nova.compute.claims [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.016193] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.234023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "162ee5e6-7f7e-406f-9495-e9849dac9420" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.234023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "162ee5e6-7f7e-406f-9495-e9849dac9420" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.543190] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.568814] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Acquiring lock "8cc47c7c-0270-4448-a0d0-c4f482118fd9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.568814] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Lock "8cc47c7c-0270-4448-a0d0-c4f482118fd9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.643414] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a58b1b-eb6b-407d-a64e-dba811801150 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.654029] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50277d61-f080-4bff-be15-d8029dbe7836 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.687160] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546817a6-e1f6-480e-a9b5-32ecd6996993 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.698085] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d462ea5-fad1-4560-8ab1-73f103351373 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.713340] env[62183]: DEBUG nova.compute.provider_tree [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.736180] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.072687] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.216527] env[62183]: DEBUG nova.scheduler.client.report [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 519.274512] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.606152] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.724207] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.182s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.724207] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 519.727152] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.184s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.730562] env[62183]: INFO nova.compute.claims [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.761314] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 520.228864] env[62183]: DEBUG nova.compute.utils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 520.232460] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 520.232771] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 520.265813] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Getting list of instances from cluster (obj){ [ 520.265813] env[62183]: value = "domain-c8" [ 520.265813] env[62183]: _type = "ClusterComputeResource" [ 520.265813] env[62183]: } {{(pid=62183) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 520.267355] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45d8037-7cc8-473b-a3a4-f9d1dd00f9b0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.283607] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Got total of 0 instances {{(pid=62183) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 520.283787] env[62183]: WARNING nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] While synchronizing instance power states, found 2 instances in the database and 0 instances on the hypervisor. [ 520.283930] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Triggering sync for uuid 53dd2793-a2da-45e4-8d25-c9b8875c0402 {{(pid=62183) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 520.284823] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Triggering sync for uuid 495347c6-1078-4a74-96b7-4ec0099cc352 {{(pid=62183) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 520.287077] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "53dd2793-a2da-45e4-8d25-c9b8875c0402" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.287334] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "495347c6-1078-4a74-96b7-4ec0099cc352" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.287526] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 520.287862] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Getting list of instances from cluster (obj){ [ 520.287862] env[62183]: value = "domain-c8" [ 520.287862] env[62183]: _type = "ClusterComputeResource" [ 520.287862] env[62183]: } {{(pid=62183) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 520.289089] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be6ec3b-cd7d-4c6e-a0a8-09a8c1f8e6a4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.299082] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Got total of 0 instances {{(pid=62183) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 520.734486] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 520.774114] env[62183]: DEBUG nova.policy [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d8b84f3981444669539bd05ba6e097f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '337f305b0ccb4cbfa2b4c0a544c6b743', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 520.856784] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7395e207-d6d1-484e-92fa-2065f2517139 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.865822] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd41bb3c-1869-4908-9b91-7170979161ab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.899511] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2093bace-8f99-4038-9c3f-91677d751723 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.908036] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d42cad9-e2aa-4614-a00d-19ed3b7ec65c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.925522] env[62183]: DEBUG nova.compute.provider_tree [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.431540] env[62183]: DEBUG nova.scheduler.client.report [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 521.747106] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 521.782407] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 521.782663] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 521.782813] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 521.783844] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 521.784244] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 521.784419] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 521.784637] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 521.784795] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 521.786855] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 521.786855] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 521.786855] env[62183]: DEBUG nova.virt.hardware [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 521.787069] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e24adb-add9-449c-81d1-000c7715a9ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.798582] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24b6583-8a6b-4d1e-be35-421d4fb05bbc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.816725] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d842b6be-6528-438a-8bea-11b3b8e48fa7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.934938] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.936215] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 521.943023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.667s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.943023] env[62183]: INFO nova.compute.claims [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.405347] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Successfully created port: d6174978-9c86-4e82-86ca-8dd5c5f3b7f1 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 522.449410] env[62183]: DEBUG nova.compute.utils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 522.451597] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 522.452838] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 522.646192] env[62183]: DEBUG nova.policy [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8f40a11608b4578a941d99daf7b37b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8781d454470f4641a3871993c292e2db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 522.956250] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.060420] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0687b6d-1b52-494e-9d2e-c326e94f8baf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.069165] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a6ea5f-8226-49d3-996a-3111ac33aa45 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.106355] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38436d59-f70b-4842-9480-6a55472693a0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.118283] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0e2139-42b7-42bd-b4b8-4093da3830ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.137434] env[62183]: DEBUG nova.compute.provider_tree [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.640387] env[62183]: DEBUG nova.scheduler.client.report [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 523.839580] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "367e93c3-7369-4baa-b854-0b73b751c378" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.839580] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "367e93c3-7369-4baa-b854-0b73b751c378" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.973522] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.001978] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.001978] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.002155] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.002397] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.002397] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.002545] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.002997] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.003849] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.003849] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.005484] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.005723] env[62183]: DEBUG nova.virt.hardware [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.006713] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b67c26-ab8d-4ec3-8bab-4e343cc6ef87 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.019896] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba2c207-4abb-4dee-9593-1ced9f8a2ce4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.071238] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Successfully created port: 8290d091-cafe-4f84-9dba-46255434140a {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 524.145990] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.205s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.146571] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 524.149590] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.544s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.150988] env[62183]: INFO nova.compute.claims [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 524.343413] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.657668] env[62183]: DEBUG nova.compute.utils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 524.659856] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 524.660557] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 524.867562] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.946407] env[62183]: DEBUG nova.policy [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0feda18d16da4f83a96d0e684d84975b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa507ce81ee94145a6f39bd7e7046abc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.159376] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.159901] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.160172] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 525.160506] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Rebuilding the list of instances to heal {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 525.166973] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 525.306262] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18f4b7d-5096-4152-a5e5-a5cbae010b73 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.316837] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6b2bde-2ce8-41c9-81f2-044650b33568 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.359992] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07cf628-6078-435e-8143-9a28e40d13b6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.368349] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc48e522-36b3-47e7-831c-c4af14727df3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.387633] env[62183]: DEBUG nova.compute.provider_tree [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.400242] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Acquiring lock "7bf2bb8b-560a-44b0-a6ac-b09398117c07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.400485] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Lock "7bf2bb8b-560a-44b0-a6ac-b09398117c07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.670500] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 525.670602] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 525.670657] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 525.670782] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 525.670931] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Didn't find any instances for network info cache update. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 525.674491] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.674878] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.679372] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.679659] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.679904] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.680132] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.680315] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 525.680465] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.891393] env[62183]: DEBUG nova.scheduler.client.report [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 525.903188] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 526.176733] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 526.183808] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.223878] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 526.228532] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 526.228532] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 526.228532] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 526.228532] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 526.228532] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 526.228532] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 526.228790] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 526.228790] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 526.229091] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 526.229429] env[62183]: DEBUG nova.virt.hardware [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 526.231038] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4daef8-82b7-448b-8ee5-3ef7904668da {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.245100] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941087e2-74ff-467e-a7b9-85e83b4c97e2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.397384] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.398056] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 526.400634] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.533s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.402578] env[62183]: INFO nova.compute.claims [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.436511] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.875800] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Successfully created port: 5e44b23d-7e3c-4b8c-92e5-7acfc0726847 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.908792] env[62183]: DEBUG nova.compute.utils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 526.916732] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 526.916956] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 527.312142] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "3f59dc4a-6781-4bc4-af96-9257d3e2eae7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.312396] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "3f59dc4a-6781-4bc4-af96-9257d3e2eae7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.363408] env[62183]: DEBUG nova.policy [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf8ea1f514f347b5a78af2385fdbdfc5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3de51b4e52024a0181c31956b4a56ab1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 527.422525] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 527.491169] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "8c9ca00f-b651-4224-bc03-9e6213d6745b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.491412] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "8c9ca00f-b651-4224-bc03-9e6213d6745b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.576700] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be888dd3-88ce-4ba4-9ef8-df95020314b9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.584728] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2d7cd5-39a7-404b-a0f0-205fe4ec4184 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.620548] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a5392b-138c-43e0-97cd-35f1d4eade5a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.628789] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4b6e9f-7eab-4019-b967-fde807cb69be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.642674] env[62183]: DEBUG nova.compute.provider_tree [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.816428] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 527.831541] env[62183]: ERROR nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1, please check neutron logs for more information. [ 527.831541] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 527.831541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.831541] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 527.831541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.831541] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 527.831541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.831541] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 527.831541] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.831541] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 527.831541] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.831541] env[62183]: ERROR nova.compute.manager raise self.value [ 527.831541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.831541] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 527.831541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.831541] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 527.832145] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.832145] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 527.832145] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1, please check neutron logs for more information. [ 527.832145] env[62183]: ERROR nova.compute.manager [ 527.832145] env[62183]: Traceback (most recent call last): [ 527.832145] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 527.832145] env[62183]: listener.cb(fileno) [ 527.832145] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.832145] env[62183]: result = function(*args, **kwargs) [ 527.832145] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 527.832145] env[62183]: return func(*args, **kwargs) [ 527.832145] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.832145] env[62183]: raise e [ 527.832145] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.832145] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 527.832145] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.832145] env[62183]: created_port_ids = self._update_ports_for_instance( [ 527.832145] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.832145] env[62183]: with excutils.save_and_reraise_exception(): [ 527.832145] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.832145] env[62183]: self.force_reraise() [ 527.832145] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.832145] env[62183]: raise self.value [ 527.832145] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.832145] env[62183]: updated_port = self._update_port( [ 527.832145] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.832145] env[62183]: _ensure_no_port_binding_failure(port) [ 527.832145] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.832145] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 527.832881] env[62183]: nova.exception.PortBindingFailed: Binding failed for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1, please check neutron logs for more information. [ 527.832881] env[62183]: Removing descriptor: 16 [ 527.832881] env[62183]: ERROR nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1, please check neutron logs for more information. [ 527.832881] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Traceback (most recent call last): [ 527.832881] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 527.832881] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] yield resources [ 527.832881] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 527.832881] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self.driver.spawn(context, instance, image_meta, [ 527.832881] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 527.832881] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.832881] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.832881] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] vm_ref = self.build_virtual_machine(instance, [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] for vif in network_info: [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] return self._sync_wrapper(fn, *args, **kwargs) [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self.wait() [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self[:] = self._gt.wait() [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] return self._exit_event.wait() [ 527.833206] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] result = hub.switch() [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] return self.greenlet.switch() [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] result = function(*args, **kwargs) [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] return func(*args, **kwargs) [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] raise e [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] nwinfo = self.network_api.allocate_for_instance( [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 527.833555] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] created_port_ids = self._update_ports_for_instance( [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] with excutils.save_and_reraise_exception(): [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self.force_reraise() [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] raise self.value [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] updated_port = self._update_port( [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] _ensure_no_port_binding_failure(port) [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.833892] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] raise exception.PortBindingFailed(port_id=port['id']) [ 527.834219] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] nova.exception.PortBindingFailed: Binding failed for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1, please check neutron logs for more information. [ 527.834219] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] [ 527.834219] env[62183]: INFO nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Terminating instance [ 527.836279] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Acquiring lock "refresh_cache-53dd2793-a2da-45e4-8d25-c9b8875c0402" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.836595] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Acquired lock "refresh_cache-53dd2793-a2da-45e4-8d25-c9b8875c0402" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.836773] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 527.994725] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.147528] env[62183]: DEBUG nova.scheduler.client.report [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 528.343217] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.431476] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.443604] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 528.479309] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 528.479538] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 528.479722] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 528.479953] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 528.480116] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 528.480260] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 528.480465] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 528.480618] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 528.480807] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 528.481475] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 528.481475] env[62183]: DEBUG nova.virt.hardware [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 528.482099] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb8d242-302e-4cdb-97e5-5f4f1d9e399e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.501120] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7cecfe2-a58d-4076-aa67-effdf6fec76f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.528833] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.560277] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Successfully created port: 4e079e3f-5e16-4c57-a80d-d7323b90da91 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 528.655975] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.255s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.656683] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 528.663508] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.478s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.663508] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.663508] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 528.663508] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.225s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.664470] env[62183]: INFO nova.compute.claims [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 528.668262] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98dcdc3d-886a-449c-862f-876ac4d7b534 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.684241] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b24af9-ed36-4fd1-b29a-6caf829442cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.708467] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee638f42-3fef-4dbd-9e2d-41600766a2da {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.717232] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582de914-98cd-441b-a4e9-b93056d4453b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.757775] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181348MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 528.757775] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.864241] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "1bfe668e-ac7b-4ac4-ae16-9642926ab598" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.864241] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "1bfe668e-ac7b-4ac4-ae16-9642926ab598" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.888556] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.169256] env[62183]: DEBUG nova.compute.utils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.171442] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 529.171575] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 529.196039] env[62183]: ERROR nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8290d091-cafe-4f84-9dba-46255434140a, please check neutron logs for more information. [ 529.196039] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.196039] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.196039] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.196039] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.196039] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.196039] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.196039] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.196039] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.196039] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 529.196039] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.196039] env[62183]: ERROR nova.compute.manager raise self.value [ 529.196039] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.196039] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.196039] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.196039] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.196720] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.196720] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.196720] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8290d091-cafe-4f84-9dba-46255434140a, please check neutron logs for more information. [ 529.196720] env[62183]: ERROR nova.compute.manager [ 529.196720] env[62183]: Traceback (most recent call last): [ 529.196720] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.196720] env[62183]: listener.cb(fileno) [ 529.196720] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.196720] env[62183]: result = function(*args, **kwargs) [ 529.196720] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.196720] env[62183]: return func(*args, **kwargs) [ 529.196720] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.196720] env[62183]: raise e [ 529.196720] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.196720] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 529.196720] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.196720] env[62183]: created_port_ids = self._update_ports_for_instance( [ 529.196720] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.196720] env[62183]: with excutils.save_and_reraise_exception(): [ 529.196720] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.196720] env[62183]: self.force_reraise() [ 529.196720] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.196720] env[62183]: raise self.value [ 529.196720] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.196720] env[62183]: updated_port = self._update_port( [ 529.196720] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.196720] env[62183]: _ensure_no_port_binding_failure(port) [ 529.196720] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.196720] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.198250] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 8290d091-cafe-4f84-9dba-46255434140a, please check neutron logs for more information. [ 529.198250] env[62183]: Removing descriptor: 14 [ 529.198250] env[62183]: ERROR nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8290d091-cafe-4f84-9dba-46255434140a, please check neutron logs for more information. [ 529.198250] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Traceback (most recent call last): [ 529.198250] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 529.198250] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] yield resources [ 529.198250] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.198250] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self.driver.spawn(context, instance, image_meta, [ 529.198250] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 529.198250] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.198250] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.198250] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] vm_ref = self.build_virtual_machine(instance, [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] for vif in network_info: [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] return self._sync_wrapper(fn, *args, **kwargs) [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self.wait() [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self[:] = self._gt.wait() [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] return self._exit_event.wait() [ 529.198738] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] result = hub.switch() [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] return self.greenlet.switch() [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] result = function(*args, **kwargs) [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] return func(*args, **kwargs) [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] raise e [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] nwinfo = self.network_api.allocate_for_instance( [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.199419] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] created_port_ids = self._update_ports_for_instance( [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] with excutils.save_and_reraise_exception(): [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self.force_reraise() [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] raise self.value [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] updated_port = self._update_port( [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] _ensure_no_port_binding_failure(port) [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.200084] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] raise exception.PortBindingFailed(port_id=port['id']) [ 529.200582] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] nova.exception.PortBindingFailed: Binding failed for port 8290d091-cafe-4f84-9dba-46255434140a, please check neutron logs for more information. [ 529.200582] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] [ 529.200582] env[62183]: INFO nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Terminating instance [ 529.200582] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Acquiring lock "refresh_cache-495347c6-1078-4a74-96b7-4ec0099cc352" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.200582] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Acquired lock "refresh_cache-495347c6-1078-4a74-96b7-4ec0099cc352" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.200582] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 529.355887] env[62183]: DEBUG nova.policy [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '970912bbb49b4d9a825afc2ef858bd03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6c01493063dd43de9f6742792e31ece1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 529.366033] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 529.393132] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Releasing lock "refresh_cache-53dd2793-a2da-45e4-8d25-c9b8875c0402" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.393132] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 529.393132] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 529.393132] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41bebbbb-b037-4861-919a-2bdedc70bbb8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.408738] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8b8e6e-e14c-44de-a5b5-9da74f600bf6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.433016] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 53dd2793-a2da-45e4-8d25-c9b8875c0402 could not be found. [ 529.433417] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 529.433883] env[62183]: INFO nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Took 0.04 seconds to destroy the instance on the hypervisor. [ 529.435343] env[62183]: DEBUG oslo.service.loopingcall [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 529.435343] env[62183]: DEBUG nova.compute.manager [-] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 529.435343] env[62183]: DEBUG nova.network.neutron [-] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 529.488542] env[62183]: DEBUG nova.network.neutron [-] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.676486] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 529.779009] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.856219] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af332cc-3cf0-45c6-9dc4-a24487e3a045 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.866154] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f9cd91-79f9-4c65-8415-1489be5e1472 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.903942] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9878b099-19e7-4ef6-9c51-c1c1c6744f74 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.916043] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1c4540-fb5a-4af7-ad5b-250aa04498bb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.921623] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.934045] env[62183]: DEBUG nova.compute.provider_tree [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.991126] env[62183]: DEBUG nova.network.neutron [-] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.122715] env[62183]: DEBUG nova.compute.manager [req-f52a3582-d0dc-47f1-aa84-afeb3970f7fe req-a195310e-e61a-4b4b-9c6f-7f64902f703f service nova] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Received event network-changed-d6174978-9c86-4e82-86ca-8dd5c5f3b7f1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.122715] env[62183]: DEBUG nova.compute.manager [req-f52a3582-d0dc-47f1-aa84-afeb3970f7fe req-a195310e-e61a-4b4b-9c6f-7f64902f703f service nova] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Refreshing instance network info cache due to event network-changed-d6174978-9c86-4e82-86ca-8dd5c5f3b7f1. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 530.122715] env[62183]: DEBUG oslo_concurrency.lockutils [req-f52a3582-d0dc-47f1-aa84-afeb3970f7fe req-a195310e-e61a-4b4b-9c6f-7f64902f703f service nova] Acquiring lock "refresh_cache-53dd2793-a2da-45e4-8d25-c9b8875c0402" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.122715] env[62183]: DEBUG oslo_concurrency.lockutils [req-f52a3582-d0dc-47f1-aa84-afeb3970f7fe req-a195310e-e61a-4b4b-9c6f-7f64902f703f service nova] Acquired lock "refresh_cache-53dd2793-a2da-45e4-8d25-c9b8875c0402" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.122715] env[62183]: DEBUG nova.network.neutron [req-f52a3582-d0dc-47f1-aa84-afeb3970f7fe req-a195310e-e61a-4b4b-9c6f-7f64902f703f service nova] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Refreshing network info cache for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 530.152231] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.436735] env[62183]: DEBUG nova.scheduler.client.report [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 530.499343] env[62183]: INFO nova.compute.manager [-] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Took 1.06 seconds to deallocate network for instance. [ 530.502043] env[62183]: DEBUG nova.compute.claims [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 530.502121] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.656089] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Releasing lock "refresh_cache-495347c6-1078-4a74-96b7-4ec0099cc352" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.656496] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 530.656596] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 530.657601] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5abbba6f-aa21-4b78-af72-2194932ab70b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.668201] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23cf76a-4c06-4859-bcf0-1677c428265f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.698610] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 530.701515] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 495347c6-1078-4a74-96b7-4ec0099cc352 could not be found. [ 530.701632] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 530.701806] env[62183]: INFO nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Took 0.05 seconds to destroy the instance on the hypervisor. [ 530.702052] env[62183]: DEBUG oslo.service.loopingcall [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 530.702492] env[62183]: DEBUG nova.compute.manager [-] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 530.702583] env[62183]: DEBUG nova.network.neutron [-] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 530.721255] env[62183]: DEBUG nova.network.neutron [req-f52a3582-d0dc-47f1-aa84-afeb3970f7fe req-a195310e-e61a-4b4b-9c6f-7f64902f703f service nova] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.733468] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.734112] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.734440] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.734731] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.735078] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.736227] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.736227] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.736227] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.736227] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.736227] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.736389] env[62183]: DEBUG nova.virt.hardware [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.737072] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ecbd1e-aea8-4849-bbda-b780a6438e73 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.741289] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Successfully created port: dd610292-8846-4e96-838a-28763bb00067 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 530.747985] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd81ce06-564e-4c84-82c8-323d2d4260c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.782020] env[62183]: DEBUG nova.network.neutron [-] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.949753] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.952427] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 530.955911] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.613s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.960280] env[62183]: INFO nova.compute.claims [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.036711] env[62183]: DEBUG nova.network.neutron [req-f52a3582-d0dc-47f1-aa84-afeb3970f7fe req-a195310e-e61a-4b4b-9c6f-7f64902f703f service nova] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.284404] env[62183]: DEBUG nova.network.neutron [-] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.470769] env[62183]: DEBUG nova.compute.utils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 531.473342] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 531.473564] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 531.543784] env[62183]: DEBUG oslo_concurrency.lockutils [req-f52a3582-d0dc-47f1-aa84-afeb3970f7fe req-a195310e-e61a-4b4b-9c6f-7f64902f703f service nova] Releasing lock "refresh_cache-53dd2793-a2da-45e4-8d25-c9b8875c0402" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.789904] env[62183]: INFO nova.compute.manager [-] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Took 1.09 seconds to deallocate network for instance. [ 531.794045] env[62183]: DEBUG nova.compute.claims [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 531.794304] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.881030] env[62183]: DEBUG nova.policy [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '44928013e21140df90dda462e4ce9db1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd550e174e35e4db0979742694b862600', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 531.978881] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.332318] env[62183]: ERROR nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847, please check neutron logs for more information. [ 532.332318] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.332318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.332318] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.332318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.332318] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.332318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.332318] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.332318] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.332318] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 532.332318] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.332318] env[62183]: ERROR nova.compute.manager raise self.value [ 532.332318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.332318] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.332318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.332318] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.333192] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.333192] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.333192] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847, please check neutron logs for more information. [ 532.333192] env[62183]: ERROR nova.compute.manager [ 532.333192] env[62183]: Traceback (most recent call last): [ 532.333192] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.333192] env[62183]: listener.cb(fileno) [ 532.333192] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.333192] env[62183]: result = function(*args, **kwargs) [ 532.333192] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 532.333192] env[62183]: return func(*args, **kwargs) [ 532.333192] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.333192] env[62183]: raise e [ 532.333192] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.333192] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 532.333192] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.333192] env[62183]: created_port_ids = self._update_ports_for_instance( [ 532.333192] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.333192] env[62183]: with excutils.save_and_reraise_exception(): [ 532.333192] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.333192] env[62183]: self.force_reraise() [ 532.333192] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.333192] env[62183]: raise self.value [ 532.333192] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.333192] env[62183]: updated_port = self._update_port( [ 532.333192] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.333192] env[62183]: _ensure_no_port_binding_failure(port) [ 532.333192] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.333192] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.334330] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847, please check neutron logs for more information. [ 532.334330] env[62183]: Removing descriptor: 17 [ 532.334330] env[62183]: ERROR nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847, please check neutron logs for more information. [ 532.334330] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Traceback (most recent call last): [ 532.334330] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 532.334330] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] yield resources [ 532.334330] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.334330] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self.driver.spawn(context, instance, image_meta, [ 532.334330] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 532.334330] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.334330] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.334330] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] vm_ref = self.build_virtual_machine(instance, [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] for vif in network_info: [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] return self._sync_wrapper(fn, *args, **kwargs) [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self.wait() [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self[:] = self._gt.wait() [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] return self._exit_event.wait() [ 532.334967] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] result = hub.switch() [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] return self.greenlet.switch() [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] result = function(*args, **kwargs) [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] return func(*args, **kwargs) [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] raise e [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] nwinfo = self.network_api.allocate_for_instance( [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.335419] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] created_port_ids = self._update_ports_for_instance( [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] with excutils.save_and_reraise_exception(): [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self.force_reraise() [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] raise self.value [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] updated_port = self._update_port( [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] _ensure_no_port_binding_failure(port) [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.335760] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] raise exception.PortBindingFailed(port_id=port['id']) [ 532.336406] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] nova.exception.PortBindingFailed: Binding failed for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847, please check neutron logs for more information. [ 532.336406] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] [ 532.336406] env[62183]: INFO nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Terminating instance [ 532.336406] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "refresh_cache-162ee5e6-7f7e-406f-9495-e9849dac9420" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.336406] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquired lock "refresh_cache-162ee5e6-7f7e-406f-9495-e9849dac9420" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.336406] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 532.383894] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a721b37-1abd-428b-9d57-0cf7135bb738 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.392432] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628b5bd8-3bb1-453e-b62a-18417c8d0f1e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.428656] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d97b5a7-0f9d-48ce-aa43-48194bbb0db0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.439155] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4502c8-f66e-4dba-9fbf-fba09d0142ba {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.453667] env[62183]: DEBUG nova.compute.provider_tree [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 532.909430] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.956885] env[62183]: DEBUG nova.scheduler.client.report [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 532.997549] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.037329] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.037616] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.037790] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.038014] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.038900] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.039478] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.039851] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.040115] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.040197] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.040412] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.040529] env[62183]: DEBUG nova.virt.hardware [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.041642] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c597106-b5d7-476a-b831-ae6c06dedead {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.059272] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca374c8d-8d38-4389-8106-b421948052a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.325588] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.386795] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Successfully created port: b77b426d-ee92-4af9-bf3a-205c34e1bf0f {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.462786] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.463389] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 533.466290] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.938s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.471058] env[62183]: INFO nova.compute.claims [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 533.776583] env[62183]: DEBUG nova.compute.manager [req-9bb57cb8-cdf5-47d3-bfc0-49472933212d req-b2c098a5-de8f-46f9-85c5-73fc4d669847 service nova] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Received event network-changed-5e44b23d-7e3c-4b8c-92e5-7acfc0726847 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.776728] env[62183]: DEBUG nova.compute.manager [req-9bb57cb8-cdf5-47d3-bfc0-49472933212d req-b2c098a5-de8f-46f9-85c5-73fc4d669847 service nova] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Refreshing instance network info cache due to event network-changed-5e44b23d-7e3c-4b8c-92e5-7acfc0726847. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 533.776928] env[62183]: DEBUG oslo_concurrency.lockutils [req-9bb57cb8-cdf5-47d3-bfc0-49472933212d req-b2c098a5-de8f-46f9-85c5-73fc4d669847 service nova] Acquiring lock "refresh_cache-162ee5e6-7f7e-406f-9495-e9849dac9420" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.781381] env[62183]: ERROR nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4e079e3f-5e16-4c57-a80d-d7323b90da91, please check neutron logs for more information. [ 533.781381] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.781381] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.781381] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.781381] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.781381] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.781381] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.781381] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.781381] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.781381] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 533.781381] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.781381] env[62183]: ERROR nova.compute.manager raise self.value [ 533.781381] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.781381] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.781381] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.781381] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.781971] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.781971] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.781971] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4e079e3f-5e16-4c57-a80d-d7323b90da91, please check neutron logs for more information. [ 533.781971] env[62183]: ERROR nova.compute.manager [ 533.781971] env[62183]: Traceback (most recent call last): [ 533.781971] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.781971] env[62183]: listener.cb(fileno) [ 533.781971] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.781971] env[62183]: result = function(*args, **kwargs) [ 533.781971] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.781971] env[62183]: return func(*args, **kwargs) [ 533.781971] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.781971] env[62183]: raise e [ 533.781971] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.781971] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 533.781971] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.781971] env[62183]: created_port_ids = self._update_ports_for_instance( [ 533.781971] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.781971] env[62183]: with excutils.save_and_reraise_exception(): [ 533.781971] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.781971] env[62183]: self.force_reraise() [ 533.781971] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.781971] env[62183]: raise self.value [ 533.781971] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.781971] env[62183]: updated_port = self._update_port( [ 533.781971] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.781971] env[62183]: _ensure_no_port_binding_failure(port) [ 533.781971] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.781971] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.783466] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 4e079e3f-5e16-4c57-a80d-d7323b90da91, please check neutron logs for more information. [ 533.783466] env[62183]: Removing descriptor: 18 [ 533.783466] env[62183]: ERROR nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4e079e3f-5e16-4c57-a80d-d7323b90da91, please check neutron logs for more information. [ 533.783466] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Traceback (most recent call last): [ 533.783466] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.783466] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] yield resources [ 533.783466] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.783466] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self.driver.spawn(context, instance, image_meta, [ 533.783466] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 533.783466] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.783466] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.783466] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] vm_ref = self.build_virtual_machine(instance, [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] for vif in network_info: [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] return self._sync_wrapper(fn, *args, **kwargs) [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self.wait() [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self[:] = self._gt.wait() [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] return self._exit_event.wait() [ 533.784901] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] result = hub.switch() [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] return self.greenlet.switch() [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] result = function(*args, **kwargs) [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] return func(*args, **kwargs) [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] raise e [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] nwinfo = self.network_api.allocate_for_instance( [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.785450] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] created_port_ids = self._update_ports_for_instance( [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] with excutils.save_and_reraise_exception(): [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self.force_reraise() [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] raise self.value [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] updated_port = self._update_port( [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] _ensure_no_port_binding_failure(port) [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.785775] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] raise exception.PortBindingFailed(port_id=port['id']) [ 533.786647] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] nova.exception.PortBindingFailed: Binding failed for port 4e079e3f-5e16-4c57-a80d-d7323b90da91, please check neutron logs for more information. [ 533.786647] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] [ 533.786647] env[62183]: INFO nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Terminating instance [ 533.789256] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Acquiring lock "refresh_cache-8cc47c7c-0270-4448-a0d0-c4f482118fd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.789494] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Acquired lock "refresh_cache-8cc47c7c-0270-4448-a0d0-c4f482118fd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.789639] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 533.832018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Releasing lock "refresh_cache-162ee5e6-7f7e-406f-9495-e9849dac9420" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.832018] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 533.832018] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 533.832018] env[62183]: DEBUG oslo_concurrency.lockutils [req-9bb57cb8-cdf5-47d3-bfc0-49472933212d req-b2c098a5-de8f-46f9-85c5-73fc4d669847 service nova] Acquired lock "refresh_cache-162ee5e6-7f7e-406f-9495-e9849dac9420" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.832018] env[62183]: DEBUG nova.network.neutron [req-9bb57cb8-cdf5-47d3-bfc0-49472933212d req-b2c098a5-de8f-46f9-85c5-73fc4d669847 service nova] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Refreshing network info cache for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 533.832289] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-446c7f76-eee0-4608-bbd0-6611f6e806eb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.844179] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b993dff-b5ec-4119-bdcf-ef372e8fe2a3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.868116] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 162ee5e6-7f7e-406f-9495-e9849dac9420 could not be found. [ 533.868400] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 533.868595] env[62183]: INFO nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Took 0.04 seconds to destroy the instance on the hypervisor. [ 533.869535] env[62183]: DEBUG oslo.service.loopingcall [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.870067] env[62183]: DEBUG nova.compute.manager [-] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 533.870067] env[62183]: DEBUG nova.network.neutron [-] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 533.898576] env[62183]: DEBUG nova.network.neutron [-] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.973707] env[62183]: DEBUG nova.compute.utils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 533.976566] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 533.977186] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 534.079498] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Acquiring lock "dda092a2-3732-44ce-953c-f6d64138c82d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.080110] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Lock "dda092a2-3732-44ce-953c-f6d64138c82d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.118465] env[62183]: DEBUG nova.policy [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '970912bbb49b4d9a825afc2ef858bd03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6c01493063dd43de9f6742792e31ece1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 534.233616] env[62183]: DEBUG nova.compute.manager [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Received event network-vif-deleted-d6174978-9c86-4e82-86ca-8dd5c5f3b7f1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.233690] env[62183]: DEBUG nova.compute.manager [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Received event network-changed-8290d091-cafe-4f84-9dba-46255434140a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.233837] env[62183]: DEBUG nova.compute.manager [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Refreshing instance network info cache due to event network-changed-8290d091-cafe-4f84-9dba-46255434140a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 534.238733] env[62183]: DEBUG oslo_concurrency.lockutils [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] Acquiring lock "refresh_cache-495347c6-1078-4a74-96b7-4ec0099cc352" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.238733] env[62183]: DEBUG oslo_concurrency.lockutils [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] Acquired lock "refresh_cache-495347c6-1078-4a74-96b7-4ec0099cc352" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.238733] env[62183]: DEBUG nova.network.neutron [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Refreshing network info cache for port 8290d091-cafe-4f84-9dba-46255434140a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 534.327120] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.383395] env[62183]: DEBUG nova.network.neutron [req-9bb57cb8-cdf5-47d3-bfc0-49472933212d req-b2c098a5-de8f-46f9-85c5-73fc4d669847 service nova] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.402155] env[62183]: DEBUG nova.network.neutron [-] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.477511] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 534.586865] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.616107] env[62183]: DEBUG nova.network.neutron [req-9bb57cb8-cdf5-47d3-bfc0-49472933212d req-b2c098a5-de8f-46f9-85c5-73fc4d669847 service nova] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.665760] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c501f37-9e78-40f2-a1f1-cc0361a0202c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.674898] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc7ca4d-801f-4def-b215-762cc7418079 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.717071] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.717071] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d927f9-4c64-4891-9afd-26b5b65a9942 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.725400] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a540a0e9-0172-4759-999e-6f65ac5c0720 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.745106] env[62183]: DEBUG nova.compute.provider_tree [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.793001] env[62183]: DEBUG nova.network.neutron [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.853125] env[62183]: ERROR nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dd610292-8846-4e96-838a-28763bb00067, please check neutron logs for more information. [ 534.853125] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.853125] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.853125] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.853125] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.853125] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.853125] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.853125] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.853125] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.853125] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 534.853125] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.853125] env[62183]: ERROR nova.compute.manager raise self.value [ 534.853125] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.853125] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.853125] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.853125] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.853690] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.853690] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.853690] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dd610292-8846-4e96-838a-28763bb00067, please check neutron logs for more information. [ 534.853690] env[62183]: ERROR nova.compute.manager [ 534.853690] env[62183]: Traceback (most recent call last): [ 534.853690] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.853690] env[62183]: listener.cb(fileno) [ 534.853690] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.853690] env[62183]: result = function(*args, **kwargs) [ 534.853690] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.853690] env[62183]: return func(*args, **kwargs) [ 534.853690] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.853690] env[62183]: raise e [ 534.853690] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.853690] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 534.853690] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.853690] env[62183]: created_port_ids = self._update_ports_for_instance( [ 534.853690] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.853690] env[62183]: with excutils.save_and_reraise_exception(): [ 534.853690] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.853690] env[62183]: self.force_reraise() [ 534.853690] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.853690] env[62183]: raise self.value [ 534.853690] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.853690] env[62183]: updated_port = self._update_port( [ 534.853690] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.853690] env[62183]: _ensure_no_port_binding_failure(port) [ 534.853690] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.853690] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.854571] env[62183]: nova.exception.PortBindingFailed: Binding failed for port dd610292-8846-4e96-838a-28763bb00067, please check neutron logs for more information. [ 534.854571] env[62183]: Removing descriptor: 16 [ 534.854571] env[62183]: ERROR nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dd610292-8846-4e96-838a-28763bb00067, please check neutron logs for more information. [ 534.854571] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Traceback (most recent call last): [ 534.854571] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.854571] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] yield resources [ 534.854571] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.854571] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self.driver.spawn(context, instance, image_meta, [ 534.854571] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 534.854571] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.854571] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.854571] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] vm_ref = self.build_virtual_machine(instance, [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] for vif in network_info: [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] return self._sync_wrapper(fn, *args, **kwargs) [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self.wait() [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self[:] = self._gt.wait() [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] return self._exit_event.wait() [ 534.854929] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] result = hub.switch() [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] return self.greenlet.switch() [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] result = function(*args, **kwargs) [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] return func(*args, **kwargs) [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] raise e [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] nwinfo = self.network_api.allocate_for_instance( [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.855306] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] created_port_ids = self._update_ports_for_instance( [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] with excutils.save_and_reraise_exception(): [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self.force_reraise() [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] raise self.value [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] updated_port = self._update_port( [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] _ensure_no_port_binding_failure(port) [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.855635] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] raise exception.PortBindingFailed(port_id=port['id']) [ 534.855922] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] nova.exception.PortBindingFailed: Binding failed for port dd610292-8846-4e96-838a-28763bb00067, please check neutron logs for more information. [ 534.855922] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] [ 534.855922] env[62183]: INFO nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Terminating instance [ 534.862334] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "refresh_cache-367e93c3-7369-4baa-b854-0b73b751c378" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.862334] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquired lock "refresh_cache-367e93c3-7369-4baa-b854-0b73b751c378" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.862334] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 534.905141] env[62183]: INFO nova.compute.manager [-] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Took 1.04 seconds to deallocate network for instance. [ 534.912943] env[62183]: DEBUG nova.compute.claims [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 534.912943] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.034679] env[62183]: DEBUG nova.network.neutron [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.119976] env[62183]: DEBUG oslo_concurrency.lockutils [req-9bb57cb8-cdf5-47d3-bfc0-49472933212d req-b2c098a5-de8f-46f9-85c5-73fc4d669847 service nova] Releasing lock "refresh_cache-162ee5e6-7f7e-406f-9495-e9849dac9420" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.126205] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.133661] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Successfully created port: f74e10c3-1b8c-4e0b-9ed4-a64a2701f514 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 535.224707] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Releasing lock "refresh_cache-8cc47c7c-0270-4448-a0d0-c4f482118fd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.225278] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.225507] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 535.225780] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-85c874b4-13cd-4d84-b096-1d38b5ed8478 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.238813] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7d0331-2b79-41c2-b8e9-12c27eef0d92 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.253361] env[62183]: DEBUG nova.scheduler.client.report [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 535.270138] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8cc47c7c-0270-4448-a0d0-c4f482118fd9 could not be found. [ 535.274183] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.275095] env[62183]: INFO nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Took 0.05 seconds to destroy the instance on the hypervisor. [ 535.275095] env[62183]: DEBUG oslo.service.loopingcall [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.275274] env[62183]: DEBUG nova.compute.manager [-] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.275371] env[62183]: DEBUG nova.network.neutron [-] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 535.323748] env[62183]: DEBUG nova.network.neutron [-] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.409657] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.489856] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 535.514879] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 535.515138] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 535.515290] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 535.515473] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 535.515609] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 535.515749] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 535.515959] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 535.516121] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 535.516279] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 535.516431] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 535.516925] env[62183]: DEBUG nova.virt.hardware [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 535.517498] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478f0de1-3a1a-4420-8d90-aa9b4f8dbe88 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.530678] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272695c2-297a-442d-8088-0ef79bb732ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.537525] env[62183]: DEBUG oslo_concurrency.lockutils [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] Releasing lock "refresh_cache-495347c6-1078-4a74-96b7-4ec0099cc352" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.537799] env[62183]: DEBUG nova.compute.manager [req-2db5d5b8-117d-4421-8450-df0fa88edc0f req-732d3b43-7980-4ad2-8e87-ef9d13841add service nova] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Received event network-vif-deleted-8290d091-cafe-4f84-9dba-46255434140a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.693990] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.759533] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.293s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.760361] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 535.762965] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 7.006s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.829899] env[62183]: DEBUG nova.network.neutron [-] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.880803] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Acquiring lock "97f8b7df-50c6-43c6-ac0d-5b61541f750b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.881054] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Lock "97f8b7df-50c6-43c6-ac0d-5b61541f750b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.200512] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Releasing lock "refresh_cache-367e93c3-7369-4baa-b854-0b73b751c378" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.201832] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 536.201832] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 536.201832] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e7ddff75-fcd1-4a02-8e3b-b3b9cbb92ca6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.212015] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdfd5f5f-d880-43d6-822e-7dde3fbe3413 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.239640] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 367e93c3-7369-4baa-b854-0b73b751c378 could not be found. [ 536.239765] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 536.239858] env[62183]: INFO nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Took 0.04 seconds to destroy the instance on the hypervisor. [ 536.240148] env[62183]: DEBUG oslo.service.loopingcall [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.241656] env[62183]: DEBUG nova.compute.manager [-] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 536.241656] env[62183]: DEBUG nova.network.neutron [-] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 536.267022] env[62183]: DEBUG nova.compute.utils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 536.274645] env[62183]: DEBUG nova.network.neutron [-] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.277657] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 536.277657] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 536.332969] env[62183]: INFO nova.compute.manager [-] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Took 1.06 seconds to deallocate network for instance. [ 536.336257] env[62183]: DEBUG nova.compute.claims [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.336586] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.368300] env[62183]: DEBUG nova.policy [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9657e30b5c04e4d95f0ab638d411469', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3331f62f874b1d90b318add34b3786', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 536.497318] env[62183]: ERROR nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f, please check neutron logs for more information. [ 536.497318] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.497318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.497318] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.497318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.497318] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.497318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.497318] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.497318] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.497318] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 536.497318] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.497318] env[62183]: ERROR nova.compute.manager raise self.value [ 536.497318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.497318] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.497318] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.497318] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.497776] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.497776] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.497776] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f, please check neutron logs for more information. [ 536.497776] env[62183]: ERROR nova.compute.manager [ 536.497776] env[62183]: Traceback (most recent call last): [ 536.497776] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.497776] env[62183]: listener.cb(fileno) [ 536.497776] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.497776] env[62183]: result = function(*args, **kwargs) [ 536.497776] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 536.497776] env[62183]: return func(*args, **kwargs) [ 536.497776] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.497776] env[62183]: raise e [ 536.497776] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.497776] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 536.497776] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.497776] env[62183]: created_port_ids = self._update_ports_for_instance( [ 536.497776] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.497776] env[62183]: with excutils.save_and_reraise_exception(): [ 536.497776] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.497776] env[62183]: self.force_reraise() [ 536.497776] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.497776] env[62183]: raise self.value [ 536.497776] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.497776] env[62183]: updated_port = self._update_port( [ 536.497776] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.497776] env[62183]: _ensure_no_port_binding_failure(port) [ 536.497776] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.497776] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.498676] env[62183]: nova.exception.PortBindingFailed: Binding failed for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f, please check neutron logs for more information. [ 536.498676] env[62183]: Removing descriptor: 14 [ 536.498676] env[62183]: ERROR nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f, please check neutron logs for more information. [ 536.498676] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Traceback (most recent call last): [ 536.498676] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 536.498676] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] yield resources [ 536.498676] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.498676] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self.driver.spawn(context, instance, image_meta, [ 536.498676] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 536.498676] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.498676] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.498676] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] vm_ref = self.build_virtual_machine(instance, [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] for vif in network_info: [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] return self._sync_wrapper(fn, *args, **kwargs) [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self.wait() [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self[:] = self._gt.wait() [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] return self._exit_event.wait() [ 536.499013] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] result = hub.switch() [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] return self.greenlet.switch() [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] result = function(*args, **kwargs) [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] return func(*args, **kwargs) [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] raise e [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] nwinfo = self.network_api.allocate_for_instance( [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.499409] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] created_port_ids = self._update_ports_for_instance( [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] with excutils.save_and_reraise_exception(): [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self.force_reraise() [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] raise self.value [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] updated_port = self._update_port( [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] _ensure_no_port_binding_failure(port) [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.499886] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] raise exception.PortBindingFailed(port_id=port['id']) [ 536.500203] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] nova.exception.PortBindingFailed: Binding failed for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f, please check neutron logs for more information. [ 536.500203] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] [ 536.500203] env[62183]: INFO nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Terminating instance [ 536.503109] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Acquiring lock "refresh_cache-7bf2bb8b-560a-44b0-a6ac-b09398117c07" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.503274] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Acquired lock "refresh_cache-7bf2bb8b-560a-44b0-a6ac-b09398117c07" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.507062] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 536.776732] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 536.784906] env[62183]: DEBUG nova.network.neutron [-] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.806695] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 53dd2793-a2da-45e4-8d25-c9b8875c0402 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 536.806842] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 495347c6-1078-4a74-96b7-4ec0099cc352 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 536.806963] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 162ee5e6-7f7e-406f-9495-e9849dac9420 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 536.807118] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 8cc47c7c-0270-4448-a0d0-c4f482118fd9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 536.807240] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 367e93c3-7369-4baa-b854-0b73b751c378 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 536.807350] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 7bf2bb8b-560a-44b0-a6ac-b09398117c07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 536.807582] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 3f59dc4a-6781-4bc4-af96-9257d3e2eae7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 536.807582] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 8c9ca00f-b651-4224-bc03-9e6213d6745b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 536.972553] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Acquiring lock "5ee79852-b725-4df7-8a64-38ad1805bb66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.972785] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Lock "5ee79852-b725-4df7-8a64-38ad1805bb66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.039074] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.169239] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Successfully created port: 421706be-65a6-4ed0-874b-2721acb0250b {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.258999] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.288939] env[62183]: INFO nova.compute.manager [-] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Took 1.05 seconds to deallocate network for instance. [ 537.292704] env[62183]: DEBUG nova.compute.claims [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 537.292892] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.315398] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 1bfe668e-ac7b-4ac4-ae16-9642926ab598 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.765314] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Releasing lock "refresh_cache-7bf2bb8b-560a-44b0-a6ac-b09398117c07" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.765685] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 537.765685] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 537.766167] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-078a783a-ee4f-40ad-8071-016e121be4ae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.777827] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b2958e-2955-48a5-8160-4ab21cc4d9ae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.793437] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 537.808911] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.808911] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.808911] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7bf2bb8b-560a-44b0-a6ac-b09398117c07 could not be found. [ 537.808911] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 537.809420] env[62183]: INFO nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Took 0.04 seconds to destroy the instance on the hypervisor. [ 537.809420] env[62183]: DEBUG oslo.service.loopingcall [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.809420] env[62183]: DEBUG nova.compute.manager [-] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 537.809550] env[62183]: DEBUG nova.network.neutron [-] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 537.824010] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance dda092a2-3732-44ce-953c-f6d64138c82d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.836532] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 537.836532] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 537.836532] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 537.836532] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 537.837793] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 537.837793] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 537.837793] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 537.837793] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 537.837793] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 537.838223] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.838223] env[62183]: DEBUG nova.virt.hardware [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.838223] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc35838-fc26-456f-b4cd-d1ac5208c320 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.847863] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2be408e-ec48-4458-89c1-365ac7cbae85 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.863499] env[62183]: DEBUG nova.network.neutron [-] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.069501] env[62183]: DEBUG nova.compute.manager [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Received event network-vif-deleted-5e44b23d-7e3c-4b8c-92e5-7acfc0726847 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.069703] env[62183]: DEBUG nova.compute.manager [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Received event network-changed-4e079e3f-5e16-4c57-a80d-d7323b90da91 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.069940] env[62183]: DEBUG nova.compute.manager [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Refreshing instance network info cache due to event network-changed-4e079e3f-5e16-4c57-a80d-d7323b90da91. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.082602] env[62183]: DEBUG oslo_concurrency.lockutils [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] Acquiring lock "refresh_cache-8cc47c7c-0270-4448-a0d0-c4f482118fd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.082602] env[62183]: DEBUG oslo_concurrency.lockutils [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] Acquired lock "refresh_cache-8cc47c7c-0270-4448-a0d0-c4f482118fd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.082602] env[62183]: DEBUG nova.network.neutron [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Refreshing network info cache for port 4e079e3f-5e16-4c57-a80d-d7323b90da91 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 538.122975] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Successfully created port: b5de4856-77ff-43f6-8972-be08c7dee15d {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 538.331146] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 97f8b7df-50c6-43c6-ac0d-5b61541f750b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.367255] env[62183]: DEBUG nova.network.neutron [-] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.646634] env[62183]: ERROR nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514, please check neutron logs for more information. [ 538.646634] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 538.646634] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.646634] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 538.646634] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.646634] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 538.646634] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.646634] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 538.646634] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.646634] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 538.646634] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.646634] env[62183]: ERROR nova.compute.manager raise self.value [ 538.646634] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.646634] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 538.646634] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.646634] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 538.647142] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.647142] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 538.647142] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514, please check neutron logs for more information. [ 538.647142] env[62183]: ERROR nova.compute.manager [ 538.647142] env[62183]: Traceback (most recent call last): [ 538.647142] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 538.647142] env[62183]: listener.cb(fileno) [ 538.647142] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.647142] env[62183]: result = function(*args, **kwargs) [ 538.647142] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 538.647142] env[62183]: return func(*args, **kwargs) [ 538.647142] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.647142] env[62183]: raise e [ 538.647142] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.647142] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 538.647142] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.647142] env[62183]: created_port_ids = self._update_ports_for_instance( [ 538.647142] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.647142] env[62183]: with excutils.save_and_reraise_exception(): [ 538.647142] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.647142] env[62183]: self.force_reraise() [ 538.647142] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.647142] env[62183]: raise self.value [ 538.647142] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.647142] env[62183]: updated_port = self._update_port( [ 538.647142] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.647142] env[62183]: _ensure_no_port_binding_failure(port) [ 538.647142] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.647142] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 538.648222] env[62183]: nova.exception.PortBindingFailed: Binding failed for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514, please check neutron logs for more information. [ 538.648222] env[62183]: Removing descriptor: 18 [ 538.648222] env[62183]: ERROR nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514, please check neutron logs for more information. [ 538.648222] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Traceback (most recent call last): [ 538.648222] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 538.648222] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] yield resources [ 538.648222] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.648222] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self.driver.spawn(context, instance, image_meta, [ 538.648222] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 538.648222] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.648222] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.648222] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] vm_ref = self.build_virtual_machine(instance, [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] for vif in network_info: [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] return self._sync_wrapper(fn, *args, **kwargs) [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self.wait() [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self[:] = self._gt.wait() [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] return self._exit_event.wait() [ 538.648651] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] result = hub.switch() [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] return self.greenlet.switch() [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] result = function(*args, **kwargs) [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] return func(*args, **kwargs) [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] raise e [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] nwinfo = self.network_api.allocate_for_instance( [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 538.649029] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] created_port_ids = self._update_ports_for_instance( [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] with excutils.save_and_reraise_exception(): [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self.force_reraise() [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] raise self.value [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] updated_port = self._update_port( [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] _ensure_no_port_binding_failure(port) [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.649379] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] raise exception.PortBindingFailed(port_id=port['id']) [ 538.649776] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] nova.exception.PortBindingFailed: Binding failed for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514, please check neutron logs for more information. [ 538.649776] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] [ 538.649776] env[62183]: INFO nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Terminating instance [ 538.653680] env[62183]: DEBUG nova.network.neutron [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.658386] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "refresh_cache-3f59dc4a-6781-4bc4-af96-9257d3e2eae7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.658726] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquired lock "refresh_cache-3f59dc4a-6781-4bc4-af96-9257d3e2eae7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.660160] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 538.680929] env[62183]: DEBUG nova.compute.manager [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Received event network-changed-dd610292-8846-4e96-838a-28763bb00067 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.680929] env[62183]: DEBUG nova.compute.manager [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Refreshing instance network info cache due to event network-changed-dd610292-8846-4e96-838a-28763bb00067. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.682573] env[62183]: DEBUG oslo_concurrency.lockutils [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] Acquiring lock "refresh_cache-367e93c3-7369-4baa-b854-0b73b751c378" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.685101] env[62183]: DEBUG oslo_concurrency.lockutils [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] Acquired lock "refresh_cache-367e93c3-7369-4baa-b854-0b73b751c378" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.685480] env[62183]: DEBUG nova.network.neutron [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Refreshing network info cache for port dd610292-8846-4e96-838a-28763bb00067 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 538.709647] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "3800f34e-b3b6-4f8f-8df9-6f7266e7f901" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.710440] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "3800f34e-b3b6-4f8f-8df9-6f7266e7f901" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.833783] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 5ee79852-b725-4df7-8a64-38ad1805bb66 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.833966] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 538.834149] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 538.872747] env[62183]: INFO nova.compute.manager [-] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Took 1.06 seconds to deallocate network for instance. [ 538.875938] env[62183]: DEBUG nova.compute.claims [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 538.876263] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.938794] env[62183]: DEBUG nova.network.neutron [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.087367] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a25926f-939c-4a5a-90de-558f24985313 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.095549] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Successfully created port: a8cd3532-44fb-43ec-9f98-55adacedc1ec {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 539.101371] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5da81f-89ab-4050-8886-098b8cbcd568 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.134519] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6f3375-3450-4a6e-a596-3c3e0d7f9bf5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.143874] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fdf0f2-7ba4-4eca-9a32-6bbe14636b7a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.164474] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.222769] env[62183]: DEBUG nova.network.neutron [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.251783] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.433974] env[62183]: DEBUG nova.network.neutron [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.441776] env[62183]: DEBUG oslo_concurrency.lockutils [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] Releasing lock "refresh_cache-8cc47c7c-0270-4448-a0d0-c4f482118fd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.442451] env[62183]: DEBUG nova.compute.manager [req-ee83d242-811c-4f21-9fc6-41af26b6068c req-6d75cbb2-2c27-495d-9516-9b81a30c48bd service nova] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Received event network-vif-deleted-4e079e3f-5e16-4c57-a80d-d7323b90da91 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.563515] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.671250] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 539.938032] env[62183]: DEBUG oslo_concurrency.lockutils [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] Releasing lock "refresh_cache-367e93c3-7369-4baa-b854-0b73b751c378" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.938032] env[62183]: DEBUG nova.compute.manager [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Received event network-vif-deleted-dd610292-8846-4e96-838a-28763bb00067 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.938032] env[62183]: DEBUG nova.compute.manager [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Received event network-changed-b77b426d-ee92-4af9-bf3a-205c34e1bf0f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.938032] env[62183]: DEBUG nova.compute.manager [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Refreshing instance network info cache due to event network-changed-b77b426d-ee92-4af9-bf3a-205c34e1bf0f. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.938032] env[62183]: DEBUG oslo_concurrency.lockutils [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] Acquiring lock "refresh_cache-7bf2bb8b-560a-44b0-a6ac-b09398117c07" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.938610] env[62183]: DEBUG oslo_concurrency.lockutils [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] Acquired lock "refresh_cache-7bf2bb8b-560a-44b0-a6ac-b09398117c07" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.938610] env[62183]: DEBUG nova.network.neutron [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Refreshing network info cache for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 540.068361] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Releasing lock "refresh_cache-3f59dc4a-6781-4bc4-af96-9257d3e2eae7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.068653] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 540.068830] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 540.069360] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5694655-e979-480c-9c50-50c8909b51c8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.081988] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bd57cb-9059-415b-a4d5-a66ea847d4c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.111365] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3f59dc4a-6781-4bc4-af96-9257d3e2eae7 could not be found. [ 540.112169] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 540.112169] env[62183]: INFO nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 540.112338] env[62183]: DEBUG oslo.service.loopingcall [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.112597] env[62183]: DEBUG nova.compute.manager [-] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.112707] env[62183]: DEBUG nova.network.neutron [-] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 540.127906] env[62183]: DEBUG nova.network.neutron [-] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.175228] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 540.175512] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.413s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.175787] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.254s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.177341] env[62183]: INFO nova.compute.claims [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 540.406346] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Acquiring lock "2aa13e44-de7e-476a-aa3d-0675d38b5ef2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.406660] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Lock "2aa13e44-de7e-476a-aa3d-0675d38b5ef2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.501573] env[62183]: DEBUG nova.network.neutron [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.631748] env[62183]: DEBUG nova.network.neutron [-] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.761025] env[62183]: DEBUG nova.network.neutron [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.134730] env[62183]: INFO nova.compute.manager [-] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Took 1.02 seconds to deallocate network for instance. [ 541.137942] env[62183]: DEBUG nova.compute.claims [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.140597] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.158022] env[62183]: DEBUG nova.compute.manager [req-33edf874-d09b-4439-95e2-1e588d6418b0 req-cd2699c2-e854-40bc-931f-ced024266d71 service nova] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Received event network-changed-f74e10c3-1b8c-4e0b-9ed4-a64a2701f514 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.158022] env[62183]: DEBUG nova.compute.manager [req-33edf874-d09b-4439-95e2-1e588d6418b0 req-cd2699c2-e854-40bc-931f-ced024266d71 service nova] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Refreshing instance network info cache due to event network-changed-f74e10c3-1b8c-4e0b-9ed4-a64a2701f514. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 541.158022] env[62183]: DEBUG oslo_concurrency.lockutils [req-33edf874-d09b-4439-95e2-1e588d6418b0 req-cd2699c2-e854-40bc-931f-ced024266d71 service nova] Acquiring lock "refresh_cache-3f59dc4a-6781-4bc4-af96-9257d3e2eae7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.158022] env[62183]: DEBUG oslo_concurrency.lockutils [req-33edf874-d09b-4439-95e2-1e588d6418b0 req-cd2699c2-e854-40bc-931f-ced024266d71 service nova] Acquired lock "refresh_cache-3f59dc4a-6781-4bc4-af96-9257d3e2eae7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.158022] env[62183]: DEBUG nova.network.neutron [req-33edf874-d09b-4439-95e2-1e588d6418b0 req-cd2699c2-e854-40bc-931f-ced024266d71 service nova] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Refreshing network info cache for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 541.263823] env[62183]: DEBUG oslo_concurrency.lockutils [req-7ba718eb-58b3-4ad8-ab6f-9c8f7560c8f8 req-d3789885-8c9a-408b-b466-e3eab736351d service nova] Releasing lock "refresh_cache-7bf2bb8b-560a-44b0-a6ac-b09398117c07" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.430080] env[62183]: ERROR nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 421706be-65a6-4ed0-874b-2721acb0250b, please check neutron logs for more information. [ 541.430080] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 541.430080] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.430080] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 541.430080] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.430080] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 541.430080] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.430080] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 541.430080] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.430080] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 541.430080] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.430080] env[62183]: ERROR nova.compute.manager raise self.value [ 541.430080] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.430080] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 541.430080] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.430080] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 541.430563] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.430563] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 541.430563] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 421706be-65a6-4ed0-874b-2721acb0250b, please check neutron logs for more information. [ 541.430563] env[62183]: ERROR nova.compute.manager [ 541.430563] env[62183]: Traceback (most recent call last): [ 541.430563] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 541.430563] env[62183]: listener.cb(fileno) [ 541.430563] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.430563] env[62183]: result = function(*args, **kwargs) [ 541.430563] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 541.430563] env[62183]: return func(*args, **kwargs) [ 541.430563] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.430563] env[62183]: raise e [ 541.430563] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.430563] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 541.430563] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.430563] env[62183]: created_port_ids = self._update_ports_for_instance( [ 541.430563] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.430563] env[62183]: with excutils.save_and_reraise_exception(): [ 541.430563] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.430563] env[62183]: self.force_reraise() [ 541.430563] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.430563] env[62183]: raise self.value [ 541.430563] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.430563] env[62183]: updated_port = self._update_port( [ 541.430563] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.430563] env[62183]: _ensure_no_port_binding_failure(port) [ 541.430563] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.430563] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 541.431358] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 421706be-65a6-4ed0-874b-2721acb0250b, please check neutron logs for more information. [ 541.431358] env[62183]: Removing descriptor: 17 [ 541.431358] env[62183]: ERROR nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 421706be-65a6-4ed0-874b-2721acb0250b, please check neutron logs for more information. [ 541.431358] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Traceback (most recent call last): [ 541.431358] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 541.431358] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] yield resources [ 541.431358] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.431358] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self.driver.spawn(context, instance, image_meta, [ 541.431358] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 541.431358] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.431358] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.431358] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] vm_ref = self.build_virtual_machine(instance, [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] for vif in network_info: [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] return self._sync_wrapper(fn, *args, **kwargs) [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self.wait() [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self[:] = self._gt.wait() [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] return self._exit_event.wait() [ 541.431727] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] result = hub.switch() [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] return self.greenlet.switch() [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] result = function(*args, **kwargs) [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] return func(*args, **kwargs) [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] raise e [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] nwinfo = self.network_api.allocate_for_instance( [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.432087] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] created_port_ids = self._update_ports_for_instance( [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] with excutils.save_and_reraise_exception(): [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self.force_reraise() [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] raise self.value [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] updated_port = self._update_port( [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] _ensure_no_port_binding_failure(port) [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.432492] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] raise exception.PortBindingFailed(port_id=port['id']) [ 541.433236] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] nova.exception.PortBindingFailed: Binding failed for port 421706be-65a6-4ed0-874b-2721acb0250b, please check neutron logs for more information. [ 541.433236] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] [ 541.433236] env[62183]: INFO nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Terminating instance [ 541.433484] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "refresh_cache-8c9ca00f-b651-4224-bc03-9e6213d6745b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.433552] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquired lock "refresh_cache-8c9ca00f-b651-4224-bc03-9e6213d6745b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.433663] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 541.486677] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49782c7-446c-493c-b5c6-8e719c5c8805 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.498019] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d8fd2b-8c20-4ce9-bf37-55f8173770aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.537973] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a194e02-c255-479c-be77-8b114fdf6f7b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.546407] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e692252-3c11-4379-8637-ab7bf9a32506 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.563313] env[62183]: DEBUG nova.compute.provider_tree [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.666406] env[62183]: DEBUG nova.compute.manager [req-25703bb8-ff87-4680-ada7-0d16cdc86ed4 req-91b091b1-149e-4656-85ea-e3180681a51c service nova] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Received event network-vif-deleted-b77b426d-ee92-4af9-bf3a-205c34e1bf0f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.692869] env[62183]: DEBUG nova.network.neutron [req-33edf874-d09b-4439-95e2-1e588d6418b0 req-cd2699c2-e854-40bc-931f-ced024266d71 service nova] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 541.988061] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.041198] env[62183]: DEBUG nova.network.neutron [req-33edf874-d09b-4439-95e2-1e588d6418b0 req-cd2699c2-e854-40bc-931f-ced024266d71 service nova] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.070939] env[62183]: DEBUG nova.scheduler.client.report [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 542.102886] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.336430] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquiring lock "a11cfd43-7516-4f98-8484-9ba2a0c5276d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.336430] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "a11cfd43-7516-4f98-8484-9ba2a0c5276d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.545293] env[62183]: DEBUG oslo_concurrency.lockutils [req-33edf874-d09b-4439-95e2-1e588d6418b0 req-cd2699c2-e854-40bc-931f-ced024266d71 service nova] Releasing lock "refresh_cache-3f59dc4a-6781-4bc4-af96-9257d3e2eae7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.545584] env[62183]: DEBUG nova.compute.manager [req-33edf874-d09b-4439-95e2-1e588d6418b0 req-cd2699c2-e854-40bc-931f-ced024266d71 service nova] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Received event network-vif-deleted-f74e10c3-1b8c-4e0b-9ed4-a64a2701f514 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.576987] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.577576] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 542.581076] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.079s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.605864] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Releasing lock "refresh_cache-8c9ca00f-b651-4224-bc03-9e6213d6745b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.606292] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 542.606489] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 542.606787] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33e79e9b-2988-47c8-ae35-38dc69e8d787 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.618304] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d14de18-86f7-48cc-a773-68da581e5f96 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.646281] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8c9ca00f-b651-4224-bc03-9e6213d6745b could not be found. [ 542.646608] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 542.646899] env[62183]: INFO nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 542.647078] env[62183]: DEBUG oslo.service.loopingcall [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.647296] env[62183]: DEBUG nova.compute.manager [-] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.647393] env[62183]: DEBUG nova.network.neutron [-] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 542.690753] env[62183]: DEBUG nova.network.neutron [-] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.094717] env[62183]: DEBUG nova.compute.utils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 543.096443] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 543.096606] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 543.249048] env[62183]: DEBUG nova.policy [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e3a11f2fdf1c4a85bf4d0f889e9356e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5805ab93ca204bafaf3f278c6bbbe3ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 543.381035] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f088057-3e85-429f-b083-4654ea5921f0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.389651] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8948f3-ae76-46d1-b5ca-f8556902e0ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.429010] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9d65d5-b66d-485f-ab0a-bd8567e42174 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.436778] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44105380-ed0a-4529-8ce7-34eefa996995 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.451825] env[62183]: DEBUG nova.compute.provider_tree [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.600937] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 543.959215] env[62183]: DEBUG nova.scheduler.client.report [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 544.293095] env[62183]: DEBUG nova.network.neutron [-] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.364645] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Successfully created port: afec792f-07f9-4601-91b2-3c602f2d176b {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 544.467021] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.467021] env[62183]: ERROR nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1, please check neutron logs for more information. [ 544.467021] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Traceback (most recent call last): [ 544.467021] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.467021] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self.driver.spawn(context, instance, image_meta, [ 544.467021] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 544.467021] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.467021] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.467021] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] vm_ref = self.build_virtual_machine(instance, [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] for vif in network_info: [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] return self._sync_wrapper(fn, *args, **kwargs) [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self.wait() [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self[:] = self._gt.wait() [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] return self._exit_event.wait() [ 544.467352] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] result = hub.switch() [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] return self.greenlet.switch() [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] result = function(*args, **kwargs) [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] return func(*args, **kwargs) [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] raise e [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] nwinfo = self.network_api.allocate_for_instance( [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 544.467746] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] created_port_ids = self._update_ports_for_instance( [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] with excutils.save_and_reraise_exception(): [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] self.force_reraise() [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] raise self.value [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] updated_port = self._update_port( [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] _ensure_no_port_binding_failure(port) [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.468109] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] raise exception.PortBindingFailed(port_id=port['id']) [ 544.468460] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] nova.exception.PortBindingFailed: Binding failed for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1, please check neutron logs for more information. [ 544.468460] env[62183]: ERROR nova.compute.manager [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] [ 544.468460] env[62183]: DEBUG nova.compute.utils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Binding failed for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 544.473097] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.676s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.481060] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Build of instance 53dd2793-a2da-45e4-8d25-c9b8875c0402 was re-scheduled: Binding failed for port d6174978-9c86-4e82-86ca-8dd5c5f3b7f1, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 544.482353] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 544.483379] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Acquiring lock "refresh_cache-53dd2793-a2da-45e4-8d25-c9b8875c0402" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.484758] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Acquired lock "refresh_cache-53dd2793-a2da-45e4-8d25-c9b8875c0402" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.484758] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 544.569702] env[62183]: DEBUG nova.compute.manager [req-23d2f1e2-dfce-451d-83e4-bd9caebbe4ba req-2908d764-ebbc-486e-8784-6e50838467ca service nova] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Received event network-changed-421706be-65a6-4ed0-874b-2721acb0250b {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.569702] env[62183]: DEBUG nova.compute.manager [req-23d2f1e2-dfce-451d-83e4-bd9caebbe4ba req-2908d764-ebbc-486e-8784-6e50838467ca service nova] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Refreshing instance network info cache due to event network-changed-421706be-65a6-4ed0-874b-2721acb0250b. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 544.570236] env[62183]: DEBUG oslo_concurrency.lockutils [req-23d2f1e2-dfce-451d-83e4-bd9caebbe4ba req-2908d764-ebbc-486e-8784-6e50838467ca service nova] Acquiring lock "refresh_cache-8c9ca00f-b651-4224-bc03-9e6213d6745b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.570236] env[62183]: DEBUG oslo_concurrency.lockutils [req-23d2f1e2-dfce-451d-83e4-bd9caebbe4ba req-2908d764-ebbc-486e-8784-6e50838467ca service nova] Acquired lock "refresh_cache-8c9ca00f-b651-4224-bc03-9e6213d6745b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.570236] env[62183]: DEBUG nova.network.neutron [req-23d2f1e2-dfce-451d-83e4-bd9caebbe4ba req-2908d764-ebbc-486e-8784-6e50838467ca service nova] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Refreshing network info cache for port 421706be-65a6-4ed0-874b-2721acb0250b {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 544.613143] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 544.637388] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.637755] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.637972] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.638374] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.638374] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.638748] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.639017] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.639241] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.639474] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.639674] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.639910] env[62183]: DEBUG nova.virt.hardware [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.641503] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b229f2f7-a9e7-49ce-860c-d23146c882ba {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.650971] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8638be28-0ff9-469d-8434-df73240a709e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.795938] env[62183]: INFO nova.compute.manager [-] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Took 2.15 seconds to deallocate network for instance. [ 544.801426] env[62183]: DEBUG nova.compute.claims [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 544.801426] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.027213] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.121496] env[62183]: DEBUG nova.network.neutron [req-23d2f1e2-dfce-451d-83e4-bd9caebbe4ba req-2908d764-ebbc-486e-8784-6e50838467ca service nova] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.134264] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "a43e24b0-c0f1-4e02-894d-0d2c8ed884d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.134680] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "a43e24b0-c0f1-4e02-894d-0d2c8ed884d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.178116] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.352225] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3175be-ab34-4b53-80cf-ea4999c2d47c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.362055] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccd3108-d5a1-403d-a7d8-72cd9dad2af6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.401761] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62cfcb4-c774-44f6-baf5-a114144b9d38 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.411707] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3cf31cb-1a6e-4f50-9f70-d44af235bceb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.433488] env[62183]: DEBUG nova.compute.provider_tree [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.596807] env[62183]: DEBUG nova.network.neutron [req-23d2f1e2-dfce-451d-83e4-bd9caebbe4ba req-2908d764-ebbc-486e-8784-6e50838467ca service nova] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.681686] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Releasing lock "refresh_cache-53dd2793-a2da-45e4-8d25-c9b8875c0402" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.681977] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 545.682113] env[62183]: DEBUG nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.682295] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 545.708346] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.937290] env[62183]: DEBUG nova.scheduler.client.report [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 546.105415] env[62183]: DEBUG oslo_concurrency.lockutils [req-23d2f1e2-dfce-451d-83e4-bd9caebbe4ba req-2908d764-ebbc-486e-8784-6e50838467ca service nova] Releasing lock "refresh_cache-8c9ca00f-b651-4224-bc03-9e6213d6745b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.105415] env[62183]: DEBUG nova.compute.manager [req-23d2f1e2-dfce-451d-83e4-bd9caebbe4ba req-2908d764-ebbc-486e-8784-6e50838467ca service nova] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Received event network-vif-deleted-421706be-65a6-4ed0-874b-2721acb0250b {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 546.214577] env[62183]: DEBUG nova.network.neutron [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.314194] env[62183]: ERROR nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port afec792f-07f9-4601-91b2-3c602f2d176b, please check neutron logs for more information. [ 546.314194] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 546.314194] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.314194] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 546.314194] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.314194] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 546.314194] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.314194] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 546.314194] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.314194] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 546.314194] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.314194] env[62183]: ERROR nova.compute.manager raise self.value [ 546.314194] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.314194] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 546.314194] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.314194] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 546.316366] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.316366] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 546.316366] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port afec792f-07f9-4601-91b2-3c602f2d176b, please check neutron logs for more information. [ 546.316366] env[62183]: ERROR nova.compute.manager [ 546.316366] env[62183]: Traceback (most recent call last): [ 546.316366] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 546.316366] env[62183]: listener.cb(fileno) [ 546.316366] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.316366] env[62183]: result = function(*args, **kwargs) [ 546.316366] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 546.316366] env[62183]: return func(*args, **kwargs) [ 546.316366] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.316366] env[62183]: raise e [ 546.316366] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.316366] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 546.316366] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.316366] env[62183]: created_port_ids = self._update_ports_for_instance( [ 546.316366] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.316366] env[62183]: with excutils.save_and_reraise_exception(): [ 546.316366] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.316366] env[62183]: self.force_reraise() [ 546.316366] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.316366] env[62183]: raise self.value [ 546.316366] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.316366] env[62183]: updated_port = self._update_port( [ 546.316366] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.316366] env[62183]: _ensure_no_port_binding_failure(port) [ 546.316366] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.316366] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 546.318578] env[62183]: nova.exception.PortBindingFailed: Binding failed for port afec792f-07f9-4601-91b2-3c602f2d176b, please check neutron logs for more information. [ 546.318578] env[62183]: Removing descriptor: 17 [ 546.318578] env[62183]: ERROR nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port afec792f-07f9-4601-91b2-3c602f2d176b, please check neutron logs for more information. [ 546.318578] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Traceback (most recent call last): [ 546.318578] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 546.318578] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] yield resources [ 546.318578] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 546.318578] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self.driver.spawn(context, instance, image_meta, [ 546.318578] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 546.318578] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.318578] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.318578] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] vm_ref = self.build_virtual_machine(instance, [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] for vif in network_info: [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] return self._sync_wrapper(fn, *args, **kwargs) [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self.wait() [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self[:] = self._gt.wait() [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] return self._exit_event.wait() [ 546.318996] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] result = hub.switch() [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] return self.greenlet.switch() [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] result = function(*args, **kwargs) [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] return func(*args, **kwargs) [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] raise e [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] nwinfo = self.network_api.allocate_for_instance( [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.319355] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] created_port_ids = self._update_ports_for_instance( [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] with excutils.save_and_reraise_exception(): [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self.force_reraise() [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] raise self.value [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] updated_port = self._update_port( [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] _ensure_no_port_binding_failure(port) [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.319688] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] raise exception.PortBindingFailed(port_id=port['id']) [ 546.320090] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] nova.exception.PortBindingFailed: Binding failed for port afec792f-07f9-4601-91b2-3c602f2d176b, please check neutron logs for more information. [ 546.320090] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] [ 546.320090] env[62183]: INFO nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Terminating instance [ 546.320090] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "refresh_cache-1bfe668e-ac7b-4ac4-ae16-9642926ab598" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.320090] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquired lock "refresh_cache-1bfe668e-ac7b-4ac4-ae16-9642926ab598" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.320090] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.450083] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.977s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.450083] env[62183]: ERROR nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8290d091-cafe-4f84-9dba-46255434140a, please check neutron logs for more information. [ 546.450083] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Traceback (most recent call last): [ 546.450083] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 546.450083] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self.driver.spawn(context, instance, image_meta, [ 546.450083] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 546.450083] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.450083] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.450083] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] vm_ref = self.build_virtual_machine(instance, [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] for vif in network_info: [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] return self._sync_wrapper(fn, *args, **kwargs) [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self.wait() [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self[:] = self._gt.wait() [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] return self._exit_event.wait() [ 546.450488] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] result = hub.switch() [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] return self.greenlet.switch() [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] result = function(*args, **kwargs) [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] return func(*args, **kwargs) [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] raise e [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] nwinfo = self.network_api.allocate_for_instance( [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 546.450806] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] created_port_ids = self._update_ports_for_instance( [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] with excutils.save_and_reraise_exception(): [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] self.force_reraise() [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] raise self.value [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] updated_port = self._update_port( [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] _ensure_no_port_binding_failure(port) [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.451126] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] raise exception.PortBindingFailed(port_id=port['id']) [ 546.451417] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] nova.exception.PortBindingFailed: Binding failed for port 8290d091-cafe-4f84-9dba-46255434140a, please check neutron logs for more information. [ 546.451417] env[62183]: ERROR nova.compute.manager [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] [ 546.451417] env[62183]: DEBUG nova.compute.utils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Binding failed for port 8290d091-cafe-4f84-9dba-46255434140a, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 546.455643] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Build of instance 495347c6-1078-4a74-96b7-4ec0099cc352 was re-scheduled: Binding failed for port 8290d091-cafe-4f84-9dba-46255434140a, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 546.455643] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 546.455643] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Acquiring lock "refresh_cache-495347c6-1078-4a74-96b7-4ec0099cc352" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.455643] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Acquired lock "refresh_cache-495347c6-1078-4a74-96b7-4ec0099cc352" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.455811] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 546.456484] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.546s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.717707] env[62183]: INFO nova.compute.manager [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] Took 1.04 seconds to deallocate network for instance. [ 546.843159] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.950917] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.976641] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.108782] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.212240] env[62183]: DEBUG nova.compute.manager [req-490b697f-680a-45da-9331-141a48f17bc7 req-43a652c2-4a87-4e00-83b5-747409f71817 service nova] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Received event network-changed-afec792f-07f9-4601-91b2-3c602f2d176b {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.212240] env[62183]: DEBUG nova.compute.manager [req-490b697f-680a-45da-9331-141a48f17bc7 req-43a652c2-4a87-4e00-83b5-747409f71817 service nova] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Refreshing instance network info cache due to event network-changed-afec792f-07f9-4601-91b2-3c602f2d176b. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 547.212240] env[62183]: DEBUG oslo_concurrency.lockutils [req-490b697f-680a-45da-9331-141a48f17bc7 req-43a652c2-4a87-4e00-83b5-747409f71817 service nova] Acquiring lock "refresh_cache-1bfe668e-ac7b-4ac4-ae16-9642926ab598" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.259852] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee9ca81-82db-48a7-8a15-f8786347ddd5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.273107] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f36f904-2ee3-42f1-a014-e84595592244 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.311508] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abde9f58-db93-4063-ad07-8c6227019e78 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.320098] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275fb346-58e0-438a-b000-cce44c59a3bb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.339154] env[62183]: DEBUG nova.compute.provider_tree [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.453860] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Releasing lock "refresh_cache-1bfe668e-ac7b-4ac4-ae16-9642926ab598" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.454274] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 547.454473] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 547.454933] env[62183]: DEBUG oslo_concurrency.lockutils [req-490b697f-680a-45da-9331-141a48f17bc7 req-43a652c2-4a87-4e00-83b5-747409f71817 service nova] Acquired lock "refresh_cache-1bfe668e-ac7b-4ac4-ae16-9642926ab598" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.455093] env[62183]: DEBUG nova.network.neutron [req-490b697f-680a-45da-9331-141a48f17bc7 req-43a652c2-4a87-4e00-83b5-747409f71817 service nova] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Refreshing network info cache for port afec792f-07f9-4601-91b2-3c602f2d176b {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 547.456044] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a4b7a82-5155-4a3a-9a0e-7c72cf57b7cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.471354] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6e2a5c-ecf0-43ba-b53f-2a4e4bde3d99 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.505536] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1bfe668e-ac7b-4ac4-ae16-9642926ab598 could not be found. [ 547.506294] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 547.506294] env[62183]: INFO nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Took 0.05 seconds to destroy the instance on the hypervisor. [ 547.506410] env[62183]: DEBUG oslo.service.loopingcall [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.506780] env[62183]: DEBUG nova.compute.manager [-] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.506845] env[62183]: DEBUG nova.network.neutron [-] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 547.527255] env[62183]: DEBUG nova.network.neutron [-] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.614470] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Releasing lock "refresh_cache-495347c6-1078-4a74-96b7-4ec0099cc352" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.614470] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 547.614603] env[62183]: DEBUG nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.614767] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 547.638093] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 547.765082] env[62183]: INFO nova.scheduler.client.report [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Deleted allocations for instance 53dd2793-a2da-45e4-8d25-c9b8875c0402 [ 547.843887] env[62183]: DEBUG nova.scheduler.client.report [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 547.995649] env[62183]: DEBUG nova.network.neutron [req-490b697f-680a-45da-9331-141a48f17bc7 req-43a652c2-4a87-4e00-83b5-747409f71817 service nova] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.029350] env[62183]: DEBUG nova.network.neutron [-] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.141730] env[62183]: DEBUG nova.network.neutron [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.209319] env[62183]: DEBUG nova.network.neutron [req-490b697f-680a-45da-9331-141a48f17bc7 req-43a652c2-4a87-4e00-83b5-747409f71817 service nova] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.276887] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6eba3b89-28fc-486d-93b8-c474e96098c0 tempest-TenantUsagesTestJSON-1861287499 tempest-TenantUsagesTestJSON-1861287499-project-member] Lock "53dd2793-a2da-45e4-8d25-c9b8875c0402" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.793s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.278553] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "53dd2793-a2da-45e4-8d25-c9b8875c0402" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 27.991s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.278553] env[62183]: INFO nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 53dd2793-a2da-45e4-8d25-c9b8875c0402] During sync_power_state the instance has a pending task (networking). Skip. [ 548.278553] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "53dd2793-a2da-45e4-8d25-c9b8875c0402" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.350841] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.894s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.351536] env[62183]: ERROR nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847, please check neutron logs for more information. [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Traceback (most recent call last): [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self.driver.spawn(context, instance, image_meta, [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] vm_ref = self.build_virtual_machine(instance, [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.351536] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] for vif in network_info: [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] return self._sync_wrapper(fn, *args, **kwargs) [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self.wait() [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self[:] = self._gt.wait() [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] return self._exit_event.wait() [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] result = hub.switch() [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.351924] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] return self.greenlet.switch() [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] result = function(*args, **kwargs) [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] return func(*args, **kwargs) [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] raise e [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] nwinfo = self.network_api.allocate_for_instance( [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] created_port_ids = self._update_ports_for_instance( [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] with excutils.save_and_reraise_exception(): [ 548.352305] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] self.force_reraise() [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] raise self.value [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] updated_port = self._update_port( [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] _ensure_no_port_binding_failure(port) [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] raise exception.PortBindingFailed(port_id=port['id']) [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] nova.exception.PortBindingFailed: Binding failed for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847, please check neutron logs for more information. [ 548.352623] env[62183]: ERROR nova.compute.manager [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] [ 548.352903] env[62183]: DEBUG nova.compute.utils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Binding failed for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 548.356190] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.230s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.358435] env[62183]: INFO nova.compute.claims [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.365019] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Build of instance 162ee5e6-7f7e-406f-9495-e9849dac9420 was re-scheduled: Binding failed for port 5e44b23d-7e3c-4b8c-92e5-7acfc0726847, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 548.365019] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 548.365019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "refresh_cache-162ee5e6-7f7e-406f-9495-e9849dac9420" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.365019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquired lock "refresh_cache-162ee5e6-7f7e-406f-9495-e9849dac9420" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.365275] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.533606] env[62183]: INFO nova.compute.manager [-] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Took 1.03 seconds to deallocate network for instance. [ 548.540073] env[62183]: DEBUG nova.compute.claims [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.540484] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.605770] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Acquiring lock "4c9534c7-5498-46a1-b6c2-c93857031078" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.606380] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Lock "4c9534c7-5498-46a1-b6c2-c93857031078" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.647616] env[62183]: INFO nova.compute.manager [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] Took 1.03 seconds to deallocate network for instance. [ 548.713622] env[62183]: DEBUG oslo_concurrency.lockutils [req-490b697f-680a-45da-9331-141a48f17bc7 req-43a652c2-4a87-4e00-83b5-747409f71817 service nova] Releasing lock "refresh_cache-1bfe668e-ac7b-4ac4-ae16-9642926ab598" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.713893] env[62183]: DEBUG nova.compute.manager [req-490b697f-680a-45da-9331-141a48f17bc7 req-43a652c2-4a87-4e00-83b5-747409f71817 service nova] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Received event network-vif-deleted-afec792f-07f9-4601-91b2-3c602f2d176b {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.784193] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.909273] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.027265] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.316509] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.535135] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Releasing lock "refresh_cache-162ee5e6-7f7e-406f-9495-e9849dac9420" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.535192] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 549.535601] env[62183]: DEBUG nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.535601] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 549.570066] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.683652] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a65d8f2-8e4b-4057-9d60-a22be34f9a7f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.693752] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9158f3da-42bb-4f05-b562-33570b22e770 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.697449] env[62183]: INFO nova.scheduler.client.report [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Deleted allocations for instance 495347c6-1078-4a74-96b7-4ec0099cc352 [ 549.732823] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7cb04d4-da7e-41db-abec-d77275c76c7a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.742863] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e309f11b-2f64-4626-8ce5-192171565c1f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.762261] env[62183]: DEBUG nova.compute.provider_tree [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.077444] env[62183]: DEBUG nova.network.neutron [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.205860] env[62183]: DEBUG oslo_concurrency.lockutils [None req-aa73d3ad-4533-49f5-837f-fda398e690c8 tempest-ImagesOneServerTestJSON-81042251 tempest-ImagesOneServerTestJSON-81042251-project-member] Lock "495347c6-1078-4a74-96b7-4ec0099cc352" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.692s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.207216] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "495347c6-1078-4a74-96b7-4ec0099cc352" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 29.920s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.207509] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-556eaae4-ba15-4db8-83ed-91dd6eb75d5f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.223579] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f674cc65-9f69-4e70-b1d6-1213ad37bf60 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.268265] env[62183]: DEBUG nova.scheduler.client.report [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 550.580659] env[62183]: INFO nova.compute.manager [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 162ee5e6-7f7e-406f-9495-e9849dac9420] Took 1.04 seconds to deallocate network for instance. [ 550.620688] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.620959] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.717451] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.753770] env[62183]: INFO nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 495347c6-1078-4a74-96b7-4ec0099cc352] During the sync_power process the instance has moved from host None to host cpu-1 [ 550.753770] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "495347c6-1078-4a74-96b7-4ec0099cc352" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.547s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.773524] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.775807] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.778258] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.442s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.243052] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.280013] env[62183]: DEBUG nova.compute.utils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.281727] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 551.281889] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 551.385033] env[62183]: DEBUG nova.policy [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'defa05b168704371926c71c1ea20d4dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d393f6921b74122b5336da1f7723873', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.578291] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f7e2d6-144a-49b9-9266-c831c1d8de16 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.587867] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ed5b27-e18b-4297-b97b-625eb0ae3772 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.629254] env[62183]: INFO nova.scheduler.client.report [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Deleted allocations for instance 162ee5e6-7f7e-406f-9495-e9849dac9420 [ 551.635952] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f08c9e-1e3e-416e-b562-bf446ddb1394 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.645973] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eaaf504-7e9b-42f2-954a-29dbe2752e47 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.662471] env[62183]: DEBUG nova.compute.provider_tree [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.786860] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.927572] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Successfully created port: f879c57d-a7ca-44cc-b53e-15c9fd762d7a {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 552.141677] env[62183]: DEBUG oslo_concurrency.lockutils [None req-087bc111-e2d0-478d-a1c7-c7062a4e2ca0 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "162ee5e6-7f7e-406f-9495-e9849dac9420" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.908s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.165907] env[62183]: DEBUG nova.scheduler.client.report [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 552.476752] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Acquiring lock "1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.477387] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Lock "1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.646900] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.674808] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.896s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.675274] env[62183]: ERROR nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4e079e3f-5e16-4c57-a80d-d7323b90da91, please check neutron logs for more information. [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Traceback (most recent call last): [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self.driver.spawn(context, instance, image_meta, [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] vm_ref = self.build_virtual_machine(instance, [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.675274] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] for vif in network_info: [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] return self._sync_wrapper(fn, *args, **kwargs) [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self.wait() [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self[:] = self._gt.wait() [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] return self._exit_event.wait() [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] result = hub.switch() [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.675660] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] return self.greenlet.switch() [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] result = function(*args, **kwargs) [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] return func(*args, **kwargs) [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] raise e [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] nwinfo = self.network_api.allocate_for_instance( [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] created_port_ids = self._update_ports_for_instance( [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] with excutils.save_and_reraise_exception(): [ 552.676053] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] self.force_reraise() [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] raise self.value [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] updated_port = self._update_port( [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] _ensure_no_port_binding_failure(port) [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] raise exception.PortBindingFailed(port_id=port['id']) [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] nova.exception.PortBindingFailed: Binding failed for port 4e079e3f-5e16-4c57-a80d-d7323b90da91, please check neutron logs for more information. [ 552.676588] env[62183]: ERROR nova.compute.manager [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] [ 552.676909] env[62183]: DEBUG nova.compute.utils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Binding failed for port 4e079e3f-5e16-4c57-a80d-d7323b90da91, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 552.677228] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.384s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.683392] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Build of instance 8cc47c7c-0270-4448-a0d0-c4f482118fd9 was re-scheduled: Binding failed for port 4e079e3f-5e16-4c57-a80d-d7323b90da91, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 552.683865] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 552.684114] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Acquiring lock "refresh_cache-8cc47c7c-0270-4448-a0d0-c4f482118fd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.684266] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Acquired lock "refresh_cache-8cc47c7c-0270-4448-a0d0-c4f482118fd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.684439] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.800419] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.831082] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.831082] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.831082] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.831365] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.831365] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.831365] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.831472] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.831805] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.832183] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.832282] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.832525] env[62183]: DEBUG nova.virt.hardware [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.833430] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ffc16c-f4f4-467d-a0b5-d3500adee045 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.842788] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eadf97b-d64d-45da-9dd4-3951413a2cc5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.181846] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.235398] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.460576] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e51d41-a0f4-4876-93e1-b92c54f07e96 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.470403] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f97ac3-e957-4195-9a77-9a04bdb45189 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.500597] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d94c542e-6fd4-416c-b412-13eb1077d589 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.510098] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85c3f88-0622-4884-8c57-480a440b4921 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.527147] env[62183]: DEBUG nova.compute.provider_tree [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.687954] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.033110] env[62183]: DEBUG nova.scheduler.client.report [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 554.189384] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Releasing lock "refresh_cache-8cc47c7c-0270-4448-a0d0-c4f482118fd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.189680] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 554.189913] env[62183]: DEBUG nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.190835] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 554.215091] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.540142] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.863s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.540883] env[62183]: ERROR nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dd610292-8846-4e96-838a-28763bb00067, please check neutron logs for more information. [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Traceback (most recent call last): [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self.driver.spawn(context, instance, image_meta, [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] vm_ref = self.build_virtual_machine(instance, [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.540883] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] for vif in network_info: [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] return self._sync_wrapper(fn, *args, **kwargs) [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self.wait() [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self[:] = self._gt.wait() [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] return self._exit_event.wait() [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] result = hub.switch() [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.541311] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] return self.greenlet.switch() [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] result = function(*args, **kwargs) [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] return func(*args, **kwargs) [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] raise e [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] nwinfo = self.network_api.allocate_for_instance( [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] created_port_ids = self._update_ports_for_instance( [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] with excutils.save_and_reraise_exception(): [ 554.541668] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] self.force_reraise() [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] raise self.value [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] updated_port = self._update_port( [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] _ensure_no_port_binding_failure(port) [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] raise exception.PortBindingFailed(port_id=port['id']) [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] nova.exception.PortBindingFailed: Binding failed for port dd610292-8846-4e96-838a-28763bb00067, please check neutron logs for more information. [ 554.542138] env[62183]: ERROR nova.compute.manager [instance: 367e93c3-7369-4baa-b854-0b73b751c378] [ 554.546100] env[62183]: DEBUG nova.compute.utils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Binding failed for port dd610292-8846-4e96-838a-28763bb00067, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 554.546100] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.667s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.549205] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Build of instance 367e93c3-7369-4baa-b854-0b73b751c378 was re-scheduled: Binding failed for port dd610292-8846-4e96-838a-28763bb00067, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 554.549205] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 554.550731] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "refresh_cache-367e93c3-7369-4baa-b854-0b73b751c378" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.550919] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquired lock "refresh_cache-367e93c3-7369-4baa-b854-0b73b751c378" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.551099] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 554.718305] env[62183]: DEBUG nova.network.neutron [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.824980] env[62183]: ERROR nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a, please check neutron logs for more information. [ 554.824980] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 554.824980] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.824980] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 554.824980] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.824980] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 554.824980] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.824980] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 554.824980] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.824980] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 554.824980] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.824980] env[62183]: ERROR nova.compute.manager raise self.value [ 554.824980] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.824980] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 554.824980] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.824980] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 554.825513] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.825513] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 554.825513] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a, please check neutron logs for more information. [ 554.825513] env[62183]: ERROR nova.compute.manager [ 554.825513] env[62183]: Traceback (most recent call last): [ 554.825513] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 554.825513] env[62183]: listener.cb(fileno) [ 554.825513] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.825513] env[62183]: result = function(*args, **kwargs) [ 554.825513] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 554.825513] env[62183]: return func(*args, **kwargs) [ 554.825513] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.825513] env[62183]: raise e [ 554.825513] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.825513] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 554.825513] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.825513] env[62183]: created_port_ids = self._update_ports_for_instance( [ 554.825513] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.825513] env[62183]: with excutils.save_and_reraise_exception(): [ 554.825513] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.825513] env[62183]: self.force_reraise() [ 554.825513] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.825513] env[62183]: raise self.value [ 554.825513] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.825513] env[62183]: updated_port = self._update_port( [ 554.825513] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.825513] env[62183]: _ensure_no_port_binding_failure(port) [ 554.825513] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.825513] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 554.826424] env[62183]: nova.exception.PortBindingFailed: Binding failed for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a, please check neutron logs for more information. [ 554.826424] env[62183]: Removing descriptor: 17 [ 554.826424] env[62183]: ERROR nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a, please check neutron logs for more information. [ 554.826424] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Traceback (most recent call last): [ 554.826424] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 554.826424] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] yield resources [ 554.826424] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.826424] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self.driver.spawn(context, instance, image_meta, [ 554.826424] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 554.826424] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.826424] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.826424] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] vm_ref = self.build_virtual_machine(instance, [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] for vif in network_info: [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] return self._sync_wrapper(fn, *args, **kwargs) [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self.wait() [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self[:] = self._gt.wait() [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] return self._exit_event.wait() [ 554.826794] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] result = hub.switch() [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] return self.greenlet.switch() [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] result = function(*args, **kwargs) [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] return func(*args, **kwargs) [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] raise e [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] nwinfo = self.network_api.allocate_for_instance( [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.827212] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] created_port_ids = self._update_ports_for_instance( [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] with excutils.save_and_reraise_exception(): [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self.force_reraise() [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] raise self.value [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] updated_port = self._update_port( [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] _ensure_no_port_binding_failure(port) [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.827612] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] raise exception.PortBindingFailed(port_id=port['id']) [ 554.827997] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] nova.exception.PortBindingFailed: Binding failed for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a, please check neutron logs for more information. [ 554.827997] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] [ 554.827997] env[62183]: INFO nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Terminating instance [ 554.828854] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Acquiring lock "refresh_cache-dda092a2-3732-44ce-953c-f6d64138c82d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.829334] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Acquired lock "refresh_cache-dda092a2-3732-44ce-953c-f6d64138c82d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.829577] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 555.111344] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.224066] env[62183]: INFO nova.compute.manager [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] [instance: 8cc47c7c-0270-4448-a0d0-c4f482118fd9] Took 1.03 seconds to deallocate network for instance. [ 555.397623] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01e1b28-b7ae-456a-96b3-e7b1a4df3a0a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.407305] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce11edf3-6686-40f1-947e-daad2d1b4d4e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.448920] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4449ca1f-f394-41b3-bb66-ef4e5f9aa39d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.458274] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c99b9ad-fc88-42d6-a1a0-66b6c6200c34 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.477480] env[62183]: DEBUG nova.compute.provider_tree [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.507247] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.630071] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.649962] env[62183]: DEBUG nova.compute.manager [req-a7830151-bdfa-4784-ab01-c8a5290236f3 req-7c18650a-4673-4181-83db-94e9a278ab0b service nova] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Received event network-changed-f879c57d-a7ca-44cc-b53e-15c9fd762d7a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 555.650184] env[62183]: DEBUG nova.compute.manager [req-a7830151-bdfa-4784-ab01-c8a5290236f3 req-7c18650a-4673-4181-83db-94e9a278ab0b service nova] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Refreshing instance network info cache due to event network-changed-f879c57d-a7ca-44cc-b53e-15c9fd762d7a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 555.650383] env[62183]: DEBUG oslo_concurrency.lockutils [req-a7830151-bdfa-4784-ab01-c8a5290236f3 req-7c18650a-4673-4181-83db-94e9a278ab0b service nova] Acquiring lock "refresh_cache-dda092a2-3732-44ce-953c-f6d64138c82d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.769868] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.980864] env[62183]: DEBUG nova.scheduler.client.report [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 556.137666] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Releasing lock "refresh_cache-367e93c3-7369-4baa-b854-0b73b751c378" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.137956] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 556.139089] env[62183]: DEBUG nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.139089] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 556.176460] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.272807] env[62183]: INFO nova.scheduler.client.report [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Deleted allocations for instance 8cc47c7c-0270-4448-a0d0-c4f482118fd9 [ 556.281317] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Releasing lock "refresh_cache-dda092a2-3732-44ce-953c-f6d64138c82d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.281745] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 556.281942] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 556.283736] env[62183]: DEBUG oslo_concurrency.lockutils [req-a7830151-bdfa-4784-ab01-c8a5290236f3 req-7c18650a-4673-4181-83db-94e9a278ab0b service nova] Acquired lock "refresh_cache-dda092a2-3732-44ce-953c-f6d64138c82d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.283919] env[62183]: DEBUG nova.network.neutron [req-a7830151-bdfa-4784-ab01-c8a5290236f3 req-7c18650a-4673-4181-83db-94e9a278ab0b service nova] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Refreshing network info cache for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 556.284978] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d98a466e-bef8-473c-85d2-a67521ff5adc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.297384] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71059e9-a99a-4370-b43d-34d0460d1588 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.325705] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dda092a2-3732-44ce-953c-f6d64138c82d could not be found. [ 556.326546] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 556.327774] env[62183]: INFO nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 556.328097] env[62183]: DEBUG oslo.service.loopingcall [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 556.328990] env[62183]: DEBUG nova.compute.manager [-] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.328990] env[62183]: DEBUG nova.network.neutron [-] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 556.363254] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Acquiring lock "3ec662c7-940e-43fa-bbea-f3bc3243adf4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.363254] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Lock "3ec662c7-940e-43fa-bbea-f3bc3243adf4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.386640] env[62183]: DEBUG nova.network.neutron [-] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.489885] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.947s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.490576] env[62183]: ERROR nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f, please check neutron logs for more information. [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Traceback (most recent call last): [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self.driver.spawn(context, instance, image_meta, [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] vm_ref = self.build_virtual_machine(instance, [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.490576] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] for vif in network_info: [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] return self._sync_wrapper(fn, *args, **kwargs) [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self.wait() [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self[:] = self._gt.wait() [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] return self._exit_event.wait() [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] result = hub.switch() [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.490946] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] return self.greenlet.switch() [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] result = function(*args, **kwargs) [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] return func(*args, **kwargs) [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] raise e [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] nwinfo = self.network_api.allocate_for_instance( [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] created_port_ids = self._update_ports_for_instance( [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] with excutils.save_and_reraise_exception(): [ 556.491464] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] self.force_reraise() [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] raise self.value [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] updated_port = self._update_port( [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] _ensure_no_port_binding_failure(port) [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] raise exception.PortBindingFailed(port_id=port['id']) [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] nova.exception.PortBindingFailed: Binding failed for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f, please check neutron logs for more information. [ 556.491869] env[62183]: ERROR nova.compute.manager [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] [ 556.492244] env[62183]: DEBUG nova.compute.utils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Binding failed for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 556.494105] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.354s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.499160] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Build of instance 7bf2bb8b-560a-44b0-a6ac-b09398117c07 was re-scheduled: Binding failed for port b77b426d-ee92-4af9-bf3a-205c34e1bf0f, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 556.499292] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 556.499518] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Acquiring lock "refresh_cache-7bf2bb8b-560a-44b0-a6ac-b09398117c07" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.499692] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Acquired lock "refresh_cache-7bf2bb8b-560a-44b0-a6ac-b09398117c07" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.499853] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 556.685768] env[62183]: DEBUG nova.network.neutron [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.793075] env[62183]: DEBUG oslo_concurrency.lockutils [None req-abdf9730-6a72-49ee-89bd-5b95213e449c tempest-FloatingIPsAssociationNegativeTestJSON-168634065 tempest-FloatingIPsAssociationNegativeTestJSON-168634065-project-member] Lock "8cc47c7c-0270-4448-a0d0-c4f482118fd9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.225s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.824142] env[62183]: DEBUG nova.network.neutron [req-a7830151-bdfa-4784-ab01-c8a5290236f3 req-7c18650a-4673-4181-83db-94e9a278ab0b service nova] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.890921] env[62183]: DEBUG nova.network.neutron [-] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.051446] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.084087] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Acquiring lock "79c10e4e-98ab-4259-8ff5-72bfa39860be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.084477] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Lock "79c10e4e-98ab-4259-8ff5-72bfa39860be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.189164] env[62183]: INFO nova.compute.manager [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 367e93c3-7369-4baa-b854-0b73b751c378] Took 1.05 seconds to deallocate network for instance. [ 557.265607] env[62183]: DEBUG nova.network.neutron [req-a7830151-bdfa-4784-ab01-c8a5290236f3 req-7c18650a-4673-4181-83db-94e9a278ab0b service nova] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.299050] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 557.307045] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.343698] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5800d95-4003-44bd-927d-8a5a9e8fbea3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.359089] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d69607d-a791-4d89-a06c-783e800f6a80 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.403370] env[62183]: INFO nova.compute.manager [-] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Took 1.07 seconds to deallocate network for instance. [ 557.406851] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3846e7-d46c-467f-9d98-a764179f2851 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.409808] env[62183]: DEBUG nova.compute.claims [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 557.409994] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.416257] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79daff29-dace-4a60-a537-04b4d931b409 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.435782] env[62183]: DEBUG nova.compute.provider_tree [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.773069] env[62183]: DEBUG oslo_concurrency.lockutils [req-a7830151-bdfa-4784-ab01-c8a5290236f3 req-7c18650a-4673-4181-83db-94e9a278ab0b service nova] Releasing lock "refresh_cache-dda092a2-3732-44ce-953c-f6d64138c82d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.812261] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Releasing lock "refresh_cache-7bf2bb8b-560a-44b0-a6ac-b09398117c07" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.812765] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 557.812765] env[62183]: DEBUG nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.812946] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.837231] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.860100] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.940442] env[62183]: DEBUG nova.scheduler.client.report [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 558.053704] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Acquiring lock "ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.053948] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Lock "ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.236521] env[62183]: INFO nova.scheduler.client.report [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Deleted allocations for instance 367e93c3-7369-4baa-b854-0b73b751c378 [ 558.334094] env[62183]: DEBUG nova.compute.manager [req-91b33e94-7018-4a43-bc32-1c2c8ae12227 req-2ff6049f-1683-4cd6-b585-71513e2793f9 service nova] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Received event network-vif-deleted-f879c57d-a7ca-44cc-b53e-15c9fd762d7a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 558.364550] env[62183]: DEBUG nova.network.neutron [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.449902] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.955s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.449902] env[62183]: ERROR nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514, please check neutron logs for more information. [ 558.449902] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Traceback (most recent call last): [ 558.449902] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 558.449902] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self.driver.spawn(context, instance, image_meta, [ 558.449902] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 558.449902] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.449902] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.449902] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] vm_ref = self.build_virtual_machine(instance, [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] for vif in network_info: [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] return self._sync_wrapper(fn, *args, **kwargs) [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self.wait() [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self[:] = self._gt.wait() [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] return self._exit_event.wait() [ 558.450297] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] result = hub.switch() [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] return self.greenlet.switch() [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] result = function(*args, **kwargs) [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] return func(*args, **kwargs) [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] raise e [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] nwinfo = self.network_api.allocate_for_instance( [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.450676] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] created_port_ids = self._update_ports_for_instance( [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] with excutils.save_and_reraise_exception(): [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] self.force_reraise() [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] raise self.value [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] updated_port = self._update_port( [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] _ensure_no_port_binding_failure(port) [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.451132] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] raise exception.PortBindingFailed(port_id=port['id']) [ 558.451480] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] nova.exception.PortBindingFailed: Binding failed for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514, please check neutron logs for more information. [ 558.451480] env[62183]: ERROR nova.compute.manager [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] [ 558.451480] env[62183]: DEBUG nova.compute.utils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Binding failed for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 558.453921] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.651s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.455871] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Build of instance 3f59dc4a-6781-4bc4-af96-9257d3e2eae7 was re-scheduled: Binding failed for port f74e10c3-1b8c-4e0b-9ed4-a64a2701f514, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 558.456343] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 558.456557] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquiring lock "refresh_cache-3f59dc4a-6781-4bc4-af96-9257d3e2eae7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.456700] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Acquired lock "refresh_cache-3f59dc4a-6781-4bc4-af96-9257d3e2eae7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.456854] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.747591] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6f2b86e4-7db9-43cf-acd5-e83196ec370d tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "367e93c3-7369-4baa-b854-0b73b751c378" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.908s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.868234] env[62183]: INFO nova.compute.manager [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] [instance: 7bf2bb8b-560a-44b0-a6ac-b09398117c07] Took 1.06 seconds to deallocate network for instance. [ 558.984248] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.130263] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.253626] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 559.367104] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a23391-ae98-48ea-ad4c-de488c362127 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.381013] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4812c125-3bbc-4df8-b74f-3c497e75462f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.417538] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b99e83-4e93-4fef-9442-d968baa73660 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.428452] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f1f8af-3c2d-4358-a520-fa302cbc9992 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.443920] env[62183]: DEBUG nova.compute.provider_tree [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.560204] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "bed9c66e-a52f-4ec9-a190-a46c92547c09" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.560264] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "bed9c66e-a52f-4ec9-a190-a46c92547c09" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.634651] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Releasing lock "refresh_cache-3f59dc4a-6781-4bc4-af96-9257d3e2eae7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.634907] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 559.635155] env[62183]: DEBUG nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.635257] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.657976] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.782603] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.920497] env[62183]: INFO nova.scheduler.client.report [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Deleted allocations for instance 7bf2bb8b-560a-44b0-a6ac-b09398117c07 [ 559.948936] env[62183]: DEBUG nova.scheduler.client.report [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 560.160764] env[62183]: DEBUG nova.network.neutron [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.432468] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5470d6d8-1c12-49ba-ab45-b493255d0d19 tempest-ImagesOneServerNegativeTestJSON-1178697849 tempest-ImagesOneServerNegativeTestJSON-1178697849-project-member] Lock "7bf2bb8b-560a-44b0-a6ac-b09398117c07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.032s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.452670] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.454661] env[62183]: ERROR nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 421706be-65a6-4ed0-874b-2721acb0250b, please check neutron logs for more information. [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Traceback (most recent call last): [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self.driver.spawn(context, instance, image_meta, [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] vm_ref = self.build_virtual_machine(instance, [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.454661] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] for vif in network_info: [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] return self._sync_wrapper(fn, *args, **kwargs) [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self.wait() [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self[:] = self._gt.wait() [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] return self._exit_event.wait() [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] result = hub.switch() [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.455747] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] return self.greenlet.switch() [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] result = function(*args, **kwargs) [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] return func(*args, **kwargs) [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] raise e [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] nwinfo = self.network_api.allocate_for_instance( [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] created_port_ids = self._update_ports_for_instance( [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] with excutils.save_and_reraise_exception(): [ 560.456669] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] self.force_reraise() [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] raise self.value [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] updated_port = self._update_port( [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] _ensure_no_port_binding_failure(port) [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] raise exception.PortBindingFailed(port_id=port['id']) [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] nova.exception.PortBindingFailed: Binding failed for port 421706be-65a6-4ed0-874b-2721acb0250b, please check neutron logs for more information. [ 560.457108] env[62183]: ERROR nova.compute.manager [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] [ 560.457476] env[62183]: DEBUG nova.compute.utils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Binding failed for port 421706be-65a6-4ed0-874b-2721acb0250b, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 560.457639] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Build of instance 8c9ca00f-b651-4224-bc03-9e6213d6745b was re-scheduled: Binding failed for port 421706be-65a6-4ed0-874b-2721acb0250b, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 560.457963] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 560.458213] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "refresh_cache-8c9ca00f-b651-4224-bc03-9e6213d6745b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.459870] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquired lock "refresh_cache-8c9ca00f-b651-4224-bc03-9e6213d6745b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.459870] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.460999] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.921s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.665043] env[62183]: INFO nova.compute.manager [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] [instance: 3f59dc4a-6781-4bc4-af96-9257d3e2eae7] Took 1.03 seconds to deallocate network for instance. [ 560.936374] env[62183]: DEBUG nova.compute.manager [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.000605] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.125398] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "650b8016-5a30-405a-b85c-6153228c9603" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.125992] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "650b8016-5a30-405a-b85c-6153228c9603" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.162222] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.397876] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb04881-6e9f-4e3a-bf71-77317aa2130f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.407342] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d13a379-8599-4dfe-944b-2e2594ed7aea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.448233] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e786d395-ffa0-4c22-9819-453087eafdbf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.459565] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd740af-5ee1-48e8-be10-097702647b1b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.472416] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.484681] env[62183]: DEBUG nova.compute.provider_tree [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.669013] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Releasing lock "refresh_cache-8c9ca00f-b651-4224-bc03-9e6213d6745b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.669751] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 561.670309] env[62183]: DEBUG nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.670309] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 561.704072] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.711996] env[62183]: INFO nova.scheduler.client.report [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Deleted allocations for instance 3f59dc4a-6781-4bc4-af96-9257d3e2eae7 [ 561.991196] env[62183]: DEBUG nova.scheduler.client.report [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 562.207270] env[62183]: DEBUG nova.network.neutron [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.223189] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0a2f7a29-ce99-4546-a7f8-cc3f66b666d0 tempest-ServersAdminTestJSON-260373778 tempest-ServersAdminTestJSON-260373778-project-member] Lock "3f59dc4a-6781-4bc4-af96-9257d3e2eae7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.910s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.496658] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.036s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.497289] env[62183]: ERROR nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port afec792f-07f9-4601-91b2-3c602f2d176b, please check neutron logs for more information. [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Traceback (most recent call last): [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self.driver.spawn(context, instance, image_meta, [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] vm_ref = self.build_virtual_machine(instance, [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.497289] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] for vif in network_info: [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] return self._sync_wrapper(fn, *args, **kwargs) [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self.wait() [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self[:] = self._gt.wait() [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] return self._exit_event.wait() [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] result = hub.switch() [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.497714] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] return self.greenlet.switch() [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] result = function(*args, **kwargs) [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] return func(*args, **kwargs) [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] raise e [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] nwinfo = self.network_api.allocate_for_instance( [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] created_port_ids = self._update_ports_for_instance( [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] with excutils.save_and_reraise_exception(): [ 562.501261] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] self.force_reraise() [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] raise self.value [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] updated_port = self._update_port( [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] _ensure_no_port_binding_failure(port) [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] raise exception.PortBindingFailed(port_id=port['id']) [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] nova.exception.PortBindingFailed: Binding failed for port afec792f-07f9-4601-91b2-3c602f2d176b, please check neutron logs for more information. [ 562.501993] env[62183]: ERROR nova.compute.manager [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] [ 562.502367] env[62183]: DEBUG nova.compute.utils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Binding failed for port afec792f-07f9-4601-91b2-3c602f2d176b, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 562.502367] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.184s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.502666] env[62183]: INFO nova.compute.claims [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 562.507357] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Build of instance 1bfe668e-ac7b-4ac4-ae16-9642926ab598 was re-scheduled: Binding failed for port afec792f-07f9-4601-91b2-3c602f2d176b, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 562.508394] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 562.508621] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "refresh_cache-1bfe668e-ac7b-4ac4-ae16-9642926ab598" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.508764] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquired lock "refresh_cache-1bfe668e-ac7b-4ac4-ae16-9642926ab598" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.508917] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.711207] env[62183]: INFO nova.compute.manager [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 8c9ca00f-b651-4224-bc03-9e6213d6745b] Took 1.04 seconds to deallocate network for instance. [ 562.726275] env[62183]: DEBUG nova.compute.manager [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.066099] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.252452] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.407374] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.753374] env[62183]: INFO nova.scheduler.client.report [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Deleted allocations for instance 8c9ca00f-b651-4224-bc03-9e6213d6745b [ 563.772435] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "cede536f-eac5-4dd0-b4b5-74cfbd24300b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.772435] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "cede536f-eac5-4dd0-b4b5-74cfbd24300b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.913074] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Releasing lock "refresh_cache-1bfe668e-ac7b-4ac4-ae16-9642926ab598" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.913304] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 563.913482] env[62183]: DEBUG nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.913711] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.919878] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25829c5d-a68c-4ba6-9a28-1d972272e3a7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.933331] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed748d6-8458-43a3-9ca1-9ebb71252a62 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.969795] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7596d95-c54e-401b-af26-fb635c078c00 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.978626] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4985788-01d1-4100-95e1-1543e03cc13a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.995188] env[62183]: DEBUG nova.compute.provider_tree [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.091129] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.273411] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d0db709-becd-4c6d-8ca4-ac51f461c830 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "8c9ca00f-b651-4224-bc03-9e6213d6745b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.782s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.499343] env[62183]: DEBUG nova.scheduler.client.report [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 564.574405] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "2c09e83a-e614-4c36-a3e1-35f6ed34e37e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.574645] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "2c09e83a-e614-4c36-a3e1-35f6ed34e37e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.594969] env[62183]: DEBUG nova.network.neutron [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.779348] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.007899] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.008594] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 565.011375] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.768s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.013549] env[62183]: INFO nova.compute.claims [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.097255] env[62183]: INFO nova.compute.manager [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 1bfe668e-ac7b-4ac4-ae16-9642926ab598] Took 1.18 seconds to deallocate network for instance. [ 565.315860] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.521298] env[62183]: DEBUG nova.compute.utils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 565.521298] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 565.521298] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 565.698966] env[62183]: DEBUG nova.policy [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'becd21431a9848a5bc434c48eb481d9d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5aa8190a7b4b48cebb379a0b30a5a186', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 566.028401] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 566.091453] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Acquiring lock "c1d8010b-14b0-4571-9d47-b0e05ebef3ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.091644] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Lock "c1d8010b-14b0-4571-9d47-b0e05ebef3ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.151810] env[62183]: INFO nova.scheduler.client.report [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Deleted allocations for instance 1bfe668e-ac7b-4ac4-ae16-9642926ab598 [ 566.394473] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad7380a-8ff4-42f9-8bbe-1df4e381b650 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.401514] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5d3f73-2049-4868-8fb9-fd148e41b4e7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.434527] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7dfbaff-a01a-4e2f-aaa6-4e4699ca85f6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.444344] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554b00e5-9f80-4206-b86c-cf141ca3cf37 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.462301] env[62183]: DEBUG nova.compute.provider_tree [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.664163] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e125df44-c18f-48b0-8f43-5321822df623 tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "1bfe668e-ac7b-4ac4-ae16-9642926ab598" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.801s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.966525] env[62183]: DEBUG nova.scheduler.client.report [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 567.041278] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 567.075846] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 567.076014] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 567.076416] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 567.076416] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 567.076502] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 567.079009] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 567.079361] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 567.080148] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 567.080148] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 567.080341] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 567.080787] env[62183]: DEBUG nova.virt.hardware [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 567.082236] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2efaea-3061-4e8d-95b8-61fd34401907 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.093211] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f04ae94-fd4d-49b8-8207-44537bc16e01 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.166903] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 567.273025] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Successfully created port: 8456e319-d388-42ab-9fb8-731185839b57 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 567.392012] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "b4a3b4f7-d726-4291-94ad-d75d656a9764" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.392713] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "b4a3b4f7-d726-4291-94ad-d75d656a9764" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.473831] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.474416] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 567.479429] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.298s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.481389] env[62183]: INFO nova.compute.claims [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 567.695440] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.989815] env[62183]: DEBUG nova.compute.utils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.995151] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.995151] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 568.205186] env[62183]: DEBUG nova.policy [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '852c3cbd24f14e6f904e82da9c0c0d1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49bb6b1a881947afb040036c4103a248', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 568.495507] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 568.839159] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-158950c5-86a9-4917-bc86-5fb81dd59120 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.848920] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f38e1f0-3bd2-4195-b216-f4d5838de413 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.883206] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74db197e-367c-4820-98c9-71f37b24b61d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.895691] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcb4f57-b941-49ac-9099-97ffe5de66ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.910910] env[62183]: DEBUG nova.compute.provider_tree [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.336961] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "8e7fbc90-5be0-4baf-b371-1a22e84dee98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.336961] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "8e7fbc90-5be0-4baf-b371-1a22e84dee98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.414534] env[62183]: DEBUG nova.scheduler.client.report [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 569.509307] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 569.539019] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:05:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1596619235',id=25,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-949106192',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 569.539019] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 569.539019] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.539253] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 569.539253] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.539511] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 569.540077] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 569.540831] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 569.540831] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 569.540831] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 569.541067] env[62183]: DEBUG nova.virt.hardware [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 569.542340] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045dbaad-7534-40ae-953b-0be01f7059f5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.546492] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Successfully created port: 6481fc64-927e-40f0-959d-2643258b26b5 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 569.557507] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07085cf4-49cb-40fd-835c-d58e16ce3a9f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.919738] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.920593] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 569.923596] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.514s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.431699] env[62183]: DEBUG nova.compute.utils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 570.443520] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 570.444174] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 570.661345] env[62183]: DEBUG nova.policy [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59edae5aed2e4ccbb980bc0973890baf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '594f9c41496146c68921dad4c8d99ec4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 570.907394] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65295b9-bf10-4bef-866f-d1446d17b826 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.922220] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae19e30-53ed-436a-9d84-71d853db2b10 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.968126] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 570.971562] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f09c30b-cf3a-4af7-b79b-17058271a540 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.981667] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fecac788-6e5e-44b3-8998-c9bb40997a7d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.999663] env[62183]: DEBUG nova.compute.provider_tree [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.388916] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "5f4edc62-4320-4aa4-ab91-ee77b757bb7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.389362] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "5f4edc62-4320-4aa4-ab91-ee77b757bb7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.425407] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.425651] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.472378] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "0120298f-dc4c-4c13-ab71-70ccd6428dc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.472644] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "0120298f-dc4c-4c13-ab71-70ccd6428dc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.506797] env[62183]: DEBUG nova.scheduler.client.report [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 571.873768] env[62183]: ERROR nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8456e319-d388-42ab-9fb8-731185839b57, please check neutron logs for more information. [ 571.873768] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 571.873768] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.873768] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 571.873768] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.873768] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 571.873768] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.873768] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 571.873768] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.873768] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 571.873768] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.873768] env[62183]: ERROR nova.compute.manager raise self.value [ 571.873768] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.873768] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 571.873768] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.873768] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 571.876498] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.876498] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 571.876498] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8456e319-d388-42ab-9fb8-731185839b57, please check neutron logs for more information. [ 571.876498] env[62183]: ERROR nova.compute.manager [ 571.876498] env[62183]: Traceback (most recent call last): [ 571.876498] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 571.876498] env[62183]: listener.cb(fileno) [ 571.876498] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.876498] env[62183]: result = function(*args, **kwargs) [ 571.876498] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 571.876498] env[62183]: return func(*args, **kwargs) [ 571.876498] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.876498] env[62183]: raise e [ 571.876498] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.876498] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 571.876498] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.876498] env[62183]: created_port_ids = self._update_ports_for_instance( [ 571.876498] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.876498] env[62183]: with excutils.save_and_reraise_exception(): [ 571.876498] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.876498] env[62183]: self.force_reraise() [ 571.876498] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.876498] env[62183]: raise self.value [ 571.876498] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.876498] env[62183]: updated_port = self._update_port( [ 571.876498] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.876498] env[62183]: _ensure_no_port_binding_failure(port) [ 571.876498] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.876498] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 571.877442] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 8456e319-d388-42ab-9fb8-731185839b57, please check neutron logs for more information. [ 571.877442] env[62183]: Removing descriptor: 17 [ 571.877442] env[62183]: ERROR nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8456e319-d388-42ab-9fb8-731185839b57, please check neutron logs for more information. [ 571.877442] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Traceback (most recent call last): [ 571.877442] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 571.877442] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] yield resources [ 571.877442] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 571.877442] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self.driver.spawn(context, instance, image_meta, [ 571.877442] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 571.877442] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.877442] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.877442] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] vm_ref = self.build_virtual_machine(instance, [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] for vif in network_info: [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] return self._sync_wrapper(fn, *args, **kwargs) [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self.wait() [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self[:] = self._gt.wait() [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] return self._exit_event.wait() [ 571.878153] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] result = hub.switch() [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] return self.greenlet.switch() [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] result = function(*args, **kwargs) [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] return func(*args, **kwargs) [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] raise e [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] nwinfo = self.network_api.allocate_for_instance( [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.878561] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] created_port_ids = self._update_ports_for_instance( [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] with excutils.save_and_reraise_exception(): [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self.force_reraise() [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] raise self.value [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] updated_port = self._update_port( [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] _ensure_no_port_binding_failure(port) [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.878940] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] raise exception.PortBindingFailed(port_id=port['id']) [ 571.881764] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] nova.exception.PortBindingFailed: Binding failed for port 8456e319-d388-42ab-9fb8-731185839b57, please check neutron logs for more information. [ 571.881764] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] [ 571.881764] env[62183]: INFO nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Terminating instance [ 571.884054] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Acquiring lock "refresh_cache-97f8b7df-50c6-43c6-ac0d-5b61541f750b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.884054] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Acquired lock "refresh_cache-97f8b7df-50c6-43c6-ac0d-5b61541f750b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.884054] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 571.982676] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.015691] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.015926] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.016095] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.016272] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.016410] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.016551] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.018884] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.019089] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.019270] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.019429] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.019602] env[62183]: DEBUG nova.virt.hardware [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.020484] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.097s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.022027] env[62183]: ERROR nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a, please check neutron logs for more information. [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Traceback (most recent call last): [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self.driver.spawn(context, instance, image_meta, [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] vm_ref = self.build_virtual_machine(instance, [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.022027] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] for vif in network_info: [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] return self._sync_wrapper(fn, *args, **kwargs) [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self.wait() [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self[:] = self._gt.wait() [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] return self._exit_event.wait() [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] result = hub.switch() [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.022667] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] return self.greenlet.switch() [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] result = function(*args, **kwargs) [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] return func(*args, **kwargs) [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] raise e [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] nwinfo = self.network_api.allocate_for_instance( [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] created_port_ids = self._update_ports_for_instance( [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] with excutils.save_and_reraise_exception(): [ 572.023109] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] self.force_reraise() [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] raise self.value [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] updated_port = self._update_port( [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] _ensure_no_port_binding_failure(port) [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] raise exception.PortBindingFailed(port_id=port['id']) [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] nova.exception.PortBindingFailed: Binding failed for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a, please check neutron logs for more information. [ 572.023520] env[62183]: ERROR nova.compute.manager [instance: dda092a2-3732-44ce-953c-f6d64138c82d] [ 572.023880] env[62183]: DEBUG nova.compute.utils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Binding failed for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.024514] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Build of instance dda092a2-3732-44ce-953c-f6d64138c82d was re-scheduled: Binding failed for port f879c57d-a7ca-44cc-b53e-15c9fd762d7a, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 572.024895] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 572.025587] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Acquiring lock "refresh_cache-dda092a2-3732-44ce-953c-f6d64138c82d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.025587] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Acquired lock "refresh_cache-dda092a2-3732-44ce-953c-f6d64138c82d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.025587] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.026958] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff7e21d-1088-4765-a68e-5cc6a234418c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.030766] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.194s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.032663] env[62183]: INFO nova.compute.claims [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.037590] env[62183]: DEBUG nova.compute.manager [req-e9dff946-43ba-4a1d-861f-b5ed3fb3b4c1 req-f2b5bed8-042a-4722-b1f4-4eadade1f564 service nova] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Received event network-changed-8456e319-d388-42ab-9fb8-731185839b57 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.037590] env[62183]: DEBUG nova.compute.manager [req-e9dff946-43ba-4a1d-861f-b5ed3fb3b4c1 req-f2b5bed8-042a-4722-b1f4-4eadade1f564 service nova] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Refreshing instance network info cache due to event network-changed-8456e319-d388-42ab-9fb8-731185839b57. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 572.037804] env[62183]: DEBUG oslo_concurrency.lockutils [req-e9dff946-43ba-4a1d-861f-b5ed3fb3b4c1 req-f2b5bed8-042a-4722-b1f4-4eadade1f564 service nova] Acquiring lock "refresh_cache-97f8b7df-50c6-43c6-ac0d-5b61541f750b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.047419] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fab498f-b2c9-4f60-b215-45a2dc5f8f2c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.188368] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Successfully created port: 98996e3e-45ac-4030-b4cb-d0490643c52c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.427123] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.585517] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.740127] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.096419] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.243243] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Releasing lock "refresh_cache-97f8b7df-50c6-43c6-ac0d-5b61541f750b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.246084] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 573.246084] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 573.246084] env[62183]: DEBUG oslo_concurrency.lockutils [req-e9dff946-43ba-4a1d-861f-b5ed3fb3b4c1 req-f2b5bed8-042a-4722-b1f4-4eadade1f564 service nova] Acquired lock "refresh_cache-97f8b7df-50c6-43c6-ac0d-5b61541f750b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.246084] env[62183]: DEBUG nova.network.neutron [req-e9dff946-43ba-4a1d-861f-b5ed3fb3b4c1 req-f2b5bed8-042a-4722-b1f4-4eadade1f564 service nova] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Refreshing network info cache for port 8456e319-d388-42ab-9fb8-731185839b57 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 573.248522] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f274a83-fe35-426f-92a3-073db50514d8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.268831] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86612db-eebe-4a10-8b4f-20230a6df10a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.311607] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 97f8b7df-50c6-43c6-ac0d-5b61541f750b could not be found. [ 573.311607] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 573.311607] env[62183]: INFO nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Took 0.07 seconds to destroy the instance on the hypervisor. [ 573.311998] env[62183]: DEBUG oslo.service.loopingcall [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 573.315364] env[62183]: DEBUG nova.compute.manager [-] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.315495] env[62183]: DEBUG nova.network.neutron [-] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.349427] env[62183]: DEBUG nova.network.neutron [-] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.447358] env[62183]: ERROR nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6481fc64-927e-40f0-959d-2643258b26b5, please check neutron logs for more information. [ 573.447358] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.447358] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.447358] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.447358] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.447358] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.447358] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.447358] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.447358] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.447358] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 573.447358] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.447358] env[62183]: ERROR nova.compute.manager raise self.value [ 573.447358] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.447358] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.447358] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.447358] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.447914] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.447914] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.447914] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6481fc64-927e-40f0-959d-2643258b26b5, please check neutron logs for more information. [ 573.447914] env[62183]: ERROR nova.compute.manager [ 573.447914] env[62183]: Traceback (most recent call last): [ 573.447914] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.447914] env[62183]: listener.cb(fileno) [ 573.447914] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.447914] env[62183]: result = function(*args, **kwargs) [ 573.447914] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 573.447914] env[62183]: return func(*args, **kwargs) [ 573.447914] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.447914] env[62183]: raise e [ 573.447914] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.447914] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 573.447914] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.447914] env[62183]: created_port_ids = self._update_ports_for_instance( [ 573.447914] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.447914] env[62183]: with excutils.save_and_reraise_exception(): [ 573.447914] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.447914] env[62183]: self.force_reraise() [ 573.447914] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.447914] env[62183]: raise self.value [ 573.447914] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.447914] env[62183]: updated_port = self._update_port( [ 573.447914] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.447914] env[62183]: _ensure_no_port_binding_failure(port) [ 573.447914] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.447914] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.448753] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 6481fc64-927e-40f0-959d-2643258b26b5, please check neutron logs for more information. [ 573.448753] env[62183]: Removing descriptor: 18 [ 573.448753] env[62183]: ERROR nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6481fc64-927e-40f0-959d-2643258b26b5, please check neutron logs for more information. [ 573.448753] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Traceback (most recent call last): [ 573.448753] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 573.448753] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] yield resources [ 573.448753] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.448753] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self.driver.spawn(context, instance, image_meta, [ 573.448753] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 573.448753] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.448753] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.448753] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] vm_ref = self.build_virtual_machine(instance, [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] for vif in network_info: [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] return self._sync_wrapper(fn, *args, **kwargs) [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self.wait() [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self[:] = self._gt.wait() [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] return self._exit_event.wait() [ 573.449147] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] result = hub.switch() [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] return self.greenlet.switch() [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] result = function(*args, **kwargs) [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] return func(*args, **kwargs) [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] raise e [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] nwinfo = self.network_api.allocate_for_instance( [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 573.449540] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] created_port_ids = self._update_ports_for_instance( [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] with excutils.save_and_reraise_exception(): [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self.force_reraise() [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] raise self.value [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] updated_port = self._update_port( [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] _ensure_no_port_binding_failure(port) [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.449957] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] raise exception.PortBindingFailed(port_id=port['id']) [ 573.450319] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] nova.exception.PortBindingFailed: Binding failed for port 6481fc64-927e-40f0-959d-2643258b26b5, please check neutron logs for more information. [ 573.450319] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] [ 573.450319] env[62183]: INFO nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Terminating instance [ 573.450983] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Acquiring lock "refresh_cache-5ee79852-b725-4df7-8a64-38ad1805bb66" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.451754] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Acquired lock "refresh_cache-5ee79852-b725-4df7-8a64-38ad1805bb66" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.451843] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 573.513908] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb00ea95-b34c-4181-a938-fa0c6a7f68fe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.523188] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a37385-db01-4d16-83b8-c6ac5e1cacfa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.568818] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbb622e-2f59-4ca7-9f02-d7b3adcddbd0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.574754] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e594983-cab7-44e8-9d43-06b666e3c8d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.593289] env[62183]: DEBUG nova.compute.provider_tree [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.601197] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Releasing lock "refresh_cache-dda092a2-3732-44ce-953c-f6d64138c82d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.602488] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 573.602488] env[62183]: DEBUG nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 573.602488] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.632786] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.779132] env[62183]: DEBUG nova.network.neutron [req-e9dff946-43ba-4a1d-861f-b5ed3fb3b4c1 req-f2b5bed8-042a-4722-b1f4-4eadade1f564 service nova] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.853689] env[62183]: DEBUG nova.network.neutron [-] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.979216] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.002850] env[62183]: DEBUG nova.network.neutron [req-e9dff946-43ba-4a1d-861f-b5ed3fb3b4c1 req-f2b5bed8-042a-4722-b1f4-4eadade1f564 service nova] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.099051] env[62183]: DEBUG nova.scheduler.client.report [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 574.134529] env[62183]: DEBUG nova.network.neutron [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.141065] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.290445] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.290559] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.358336] env[62183]: INFO nova.compute.manager [-] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Took 1.04 seconds to deallocate network for instance. [ 574.361087] env[62183]: DEBUG nova.compute.claims [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 574.361328] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.505701] env[62183]: DEBUG oslo_concurrency.lockutils [req-e9dff946-43ba-4a1d-861f-b5ed3fb3b4c1 req-f2b5bed8-042a-4722-b1f4-4eadade1f564 service nova] Releasing lock "refresh_cache-97f8b7df-50c6-43c6-ac0d-5b61541f750b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.604417] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.604940] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 574.612104] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.826s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.612104] env[62183]: INFO nova.compute.claims [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 574.639146] env[62183]: INFO nova.compute.manager [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] [instance: dda092a2-3732-44ce-953c-f6d64138c82d] Took 1.04 seconds to deallocate network for instance. [ 574.645142] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Releasing lock "refresh_cache-5ee79852-b725-4df7-8a64-38ad1805bb66" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.645577] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 574.645966] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 574.646060] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c61d1a12-1633-49aa-a195-22d581cfa6be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.659368] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb65dd8-950a-4df5-bd9e-00d63961da50 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.695054] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ee79852-b725-4df7-8a64-38ad1805bb66 could not be found. [ 574.695346] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.695579] env[62183]: INFO nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Took 0.05 seconds to destroy the instance on the hypervisor. [ 574.695867] env[62183]: DEBUG oslo.service.loopingcall [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.696157] env[62183]: DEBUG nova.compute.manager [-] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.696278] env[62183]: DEBUG nova.network.neutron [-] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.713528] env[62183]: ERROR nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 98996e3e-45ac-4030-b4cb-d0490643c52c, please check neutron logs for more information. [ 574.713528] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 574.713528] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.713528] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 574.713528] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 574.713528] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 574.713528] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 574.713528] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 574.713528] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.713528] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 574.713528] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.713528] env[62183]: ERROR nova.compute.manager raise self.value [ 574.713528] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 574.713528] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 574.713528] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.713528] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 574.714108] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.714108] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 574.714108] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 98996e3e-45ac-4030-b4cb-d0490643c52c, please check neutron logs for more information. [ 574.714108] env[62183]: ERROR nova.compute.manager [ 574.714108] env[62183]: Traceback (most recent call last): [ 574.714108] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 574.714108] env[62183]: listener.cb(fileno) [ 574.714108] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.714108] env[62183]: result = function(*args, **kwargs) [ 574.714108] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 574.714108] env[62183]: return func(*args, **kwargs) [ 574.714108] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.714108] env[62183]: raise e [ 574.714108] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.714108] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 574.714108] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 574.714108] env[62183]: created_port_ids = self._update_ports_for_instance( [ 574.714108] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 574.714108] env[62183]: with excutils.save_and_reraise_exception(): [ 574.714108] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.714108] env[62183]: self.force_reraise() [ 574.714108] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.714108] env[62183]: raise self.value [ 574.714108] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 574.714108] env[62183]: updated_port = self._update_port( [ 574.714108] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.714108] env[62183]: _ensure_no_port_binding_failure(port) [ 574.714108] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.714108] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 574.714901] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 98996e3e-45ac-4030-b4cb-d0490643c52c, please check neutron logs for more information. [ 574.714901] env[62183]: Removing descriptor: 14 [ 574.714901] env[62183]: ERROR nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 98996e3e-45ac-4030-b4cb-d0490643c52c, please check neutron logs for more information. [ 574.714901] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Traceback (most recent call last): [ 574.714901] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 574.714901] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] yield resources [ 574.714901] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 574.714901] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self.driver.spawn(context, instance, image_meta, [ 574.714901] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 574.714901] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.714901] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.714901] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] vm_ref = self.build_virtual_machine(instance, [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] for vif in network_info: [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] return self._sync_wrapper(fn, *args, **kwargs) [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self.wait() [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self[:] = self._gt.wait() [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] return self._exit_event.wait() [ 574.715380] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] result = hub.switch() [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] return self.greenlet.switch() [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] result = function(*args, **kwargs) [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] return func(*args, **kwargs) [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] raise e [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] nwinfo = self.network_api.allocate_for_instance( [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 574.715819] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] created_port_ids = self._update_ports_for_instance( [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] with excutils.save_and_reraise_exception(): [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self.force_reraise() [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] raise self.value [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] updated_port = self._update_port( [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] _ensure_no_port_binding_failure(port) [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.716188] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] raise exception.PortBindingFailed(port_id=port['id']) [ 574.716581] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] nova.exception.PortBindingFailed: Binding failed for port 98996e3e-45ac-4030-b4cb-d0490643c52c, please check neutron logs for more information. [ 574.716581] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] [ 574.716581] env[62183]: INFO nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Terminating instance [ 574.721894] env[62183]: DEBUG nova.network.neutron [-] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.722134] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.722336] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.722510] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.828605] env[62183]: DEBUG nova.compute.manager [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Received event network-vif-deleted-8456e319-d388-42ab-9fb8-731185839b57 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.829170] env[62183]: DEBUG nova.compute.manager [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Received event network-changed-6481fc64-927e-40f0-959d-2643258b26b5 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.829170] env[62183]: DEBUG nova.compute.manager [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Refreshing instance network info cache due to event network-changed-6481fc64-927e-40f0-959d-2643258b26b5. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 574.829170] env[62183]: DEBUG oslo_concurrency.lockutils [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] Acquiring lock "refresh_cache-5ee79852-b725-4df7-8a64-38ad1805bb66" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.829334] env[62183]: DEBUG oslo_concurrency.lockutils [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] Acquired lock "refresh_cache-5ee79852-b725-4df7-8a64-38ad1805bb66" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.829648] env[62183]: DEBUG nova.network.neutron [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Refreshing network info cache for port 6481fc64-927e-40f0-959d-2643258b26b5 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 574.983955] env[62183]: DEBUG nova.compute.manager [req-30f1e099-b90f-4cc3-ab86-4bca84cd2ee3 req-8d61520f-b6fd-4468-9f83-38072b2854e3 service nova] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Received event network-changed-98996e3e-45ac-4030-b4cb-d0490643c52c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.984198] env[62183]: DEBUG nova.compute.manager [req-30f1e099-b90f-4cc3-ab86-4bca84cd2ee3 req-8d61520f-b6fd-4468-9f83-38072b2854e3 service nova] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Refreshing instance network info cache due to event network-changed-98996e3e-45ac-4030-b4cb-d0490643c52c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 574.987021] env[62183]: DEBUG oslo_concurrency.lockutils [req-30f1e099-b90f-4cc3-ab86-4bca84cd2ee3 req-8d61520f-b6fd-4468-9f83-38072b2854e3 service nova] Acquiring lock "refresh_cache-01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.122019] env[62183]: DEBUG nova.compute.utils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.123088] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.123281] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 575.233479] env[62183]: DEBUG nova.network.neutron [-] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.240196] env[62183]: DEBUG nova.policy [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '20aeadc9abc1467eac21bd9ded19356a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd498301db0a947f588b0f299efc9a9db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 575.287805] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.358106] env[62183]: DEBUG nova.network.neutron [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.469057] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Acquiring lock "f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.469731] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Lock "f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.623593] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 575.633173] env[62183]: DEBUG nova.network.neutron [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.636389] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.680221] env[62183]: INFO nova.scheduler.client.report [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Deleted allocations for instance dda092a2-3732-44ce-953c-f6d64138c82d [ 575.739035] env[62183]: INFO nova.compute.manager [-] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Took 1.04 seconds to deallocate network for instance. [ 575.749812] env[62183]: DEBUG nova.compute.claims [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.750055] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.072469] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "ee99cd66-908e-448f-9cce-ebe78a32214d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.072714] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "ee99cd66-908e-448f-9cce-ebe78a32214d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.137236] env[62183]: DEBUG oslo_concurrency.lockutils [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] Releasing lock "refresh_cache-5ee79852-b725-4df7-8a64-38ad1805bb66" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.137236] env[62183]: DEBUG nova.compute.manager [req-0abd732a-877a-4ebd-b71e-775daf531943 req-cf957516-11fc-4ec4-8aee-77815b13819e service nova] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Received event network-vif-deleted-6481fc64-927e-40f0-959d-2643258b26b5 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.137512] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.137968] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 576.138403] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 576.138796] env[62183]: DEBUG oslo_concurrency.lockutils [req-30f1e099-b90f-4cc3-ab86-4bca84cd2ee3 req-8d61520f-b6fd-4468-9f83-38072b2854e3 service nova] Acquired lock "refresh_cache-01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.139047] env[62183]: DEBUG nova.network.neutron [req-30f1e099-b90f-4cc3-ab86-4bca84cd2ee3 req-8d61520f-b6fd-4468-9f83-38072b2854e3 service nova] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Refreshing network info cache for port 98996e3e-45ac-4030-b4cb-d0490643c52c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 576.140243] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31e16cf1-5018-4968-aa99-8f9a9364eeab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.154903] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-315e9c65-eae6-47d8-a972-3ebba3ee05ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.189032] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff could not be found. [ 576.189825] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 576.189825] env[62183]: INFO nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Took 0.05 seconds to destroy the instance on the hypervisor. [ 576.189825] env[62183]: DEBUG oslo.service.loopingcall [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.192478] env[62183]: DEBUG oslo_concurrency.lockutils [None req-035ebf0b-e5bb-4b4b-ab6c-fd776c0e26c7 tempest-ServerAddressesNegativeTestJSON-664971175 tempest-ServerAddressesNegativeTestJSON-664971175-project-member] Lock "dda092a2-3732-44ce-953c-f6d64138c82d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.113s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.192602] env[62183]: DEBUG nova.compute.manager [-] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.193268] env[62183]: DEBUG nova.network.neutron [-] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.196611] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c095e29-2048-452c-97aa-709fec8ae78e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.204946] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad28ec6e-feba-4d7a-b7e4-b98eefa75f98 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.243673] env[62183]: DEBUG nova.network.neutron [-] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.246955] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c76b52-833c-47e2-aa11-715dcc55c62a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.256550] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6999eb62-5a3c-422d-99be-cf3bbd87e8f3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.274134] env[62183]: DEBUG nova.compute.provider_tree [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.480558] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Successfully created port: ba424952-b1fa-476b-9694-d0e9b644f101 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.639955] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 576.666674] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 576.667439] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 576.667439] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.667630] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 576.668063] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.668279] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 576.668759] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 576.668759] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 576.669316] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 576.669316] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 576.669316] env[62183]: DEBUG nova.virt.hardware [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 576.670088] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-453e164a-4e35-46a2-9007-b51fc2c8acad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.679021] env[62183]: DEBUG nova.network.neutron [req-30f1e099-b90f-4cc3-ab86-4bca84cd2ee3 req-8d61520f-b6fd-4468-9f83-38072b2854e3 service nova] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.684120] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08615c54-a755-498e-b5a4-d3fae5c86ae8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.700584] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 576.750176] env[62183]: DEBUG nova.network.neutron [-] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.776848] env[62183]: DEBUG nova.scheduler.client.report [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 576.874966] env[62183]: DEBUG nova.network.neutron [req-30f1e099-b90f-4cc3-ab86-4bca84cd2ee3 req-8d61520f-b6fd-4468-9f83-38072b2854e3 service nova] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.231395] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.257885] env[62183]: INFO nova.compute.manager [-] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Took 1.06 seconds to deallocate network for instance. [ 577.260738] env[62183]: DEBUG nova.compute.claims [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 577.260738] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.284119] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.676s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.286789] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.292360] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.820s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.293918] env[62183]: INFO nova.compute.claims [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.378333] env[62183]: DEBUG oslo_concurrency.lockutils [req-30f1e099-b90f-4cc3-ab86-4bca84cd2ee3 req-8d61520f-b6fd-4468-9f83-38072b2854e3 service nova] Releasing lock "refresh_cache-01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.562306] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Acquiring lock "14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.562608] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Lock "14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.641547] env[62183]: DEBUG nova.compute.manager [req-622d1d9b-2db3-45fd-9c99-9ca59ff04aad req-ef707d97-4ffc-4fb5-94a6-fe9a7f3720ac service nova] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Received event network-vif-deleted-98996e3e-45ac-4030-b4cb-d0490643c52c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.799808] env[62183]: DEBUG nova.compute.utils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 577.803479] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 577.803479] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 577.882548] env[62183]: DEBUG nova.policy [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ae9dd0f01244b77b7d680d54ffe5302', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be4c0f847d554c0ea179bb51ccb37f33', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 577.896123] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "4af88a37-5aa2-47af-9dd9-8233b1bbf077" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.896486] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "4af88a37-5aa2-47af-9dd9-8233b1bbf077" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.066542] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquiring lock "8397279a-7a43-4fb0-bc27-17e17e258527" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.066772] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "8397279a-7a43-4fb0-bc27-17e17e258527" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.226189] env[62183]: ERROR nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ba424952-b1fa-476b-9694-d0e9b644f101, please check neutron logs for more information. [ 578.226189] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 578.226189] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.226189] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 578.226189] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.226189] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 578.226189] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.226189] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 578.226189] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.226189] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 578.226189] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.226189] env[62183]: ERROR nova.compute.manager raise self.value [ 578.226189] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.226189] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 578.226189] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.226189] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 578.226664] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.226664] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 578.226664] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ba424952-b1fa-476b-9694-d0e9b644f101, please check neutron logs for more information. [ 578.226664] env[62183]: ERROR nova.compute.manager [ 578.226664] env[62183]: Traceback (most recent call last): [ 578.226664] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 578.226664] env[62183]: listener.cb(fileno) [ 578.226664] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.226664] env[62183]: result = function(*args, **kwargs) [ 578.226664] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.226664] env[62183]: return func(*args, **kwargs) [ 578.226664] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.226664] env[62183]: raise e [ 578.226664] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.226664] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 578.226664] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.226664] env[62183]: created_port_ids = self._update_ports_for_instance( [ 578.226664] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.226664] env[62183]: with excutils.save_and_reraise_exception(): [ 578.226664] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.226664] env[62183]: self.force_reraise() [ 578.226664] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.226664] env[62183]: raise self.value [ 578.226664] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.226664] env[62183]: updated_port = self._update_port( [ 578.226664] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.226664] env[62183]: _ensure_no_port_binding_failure(port) [ 578.226664] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.226664] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 578.227573] env[62183]: nova.exception.PortBindingFailed: Binding failed for port ba424952-b1fa-476b-9694-d0e9b644f101, please check neutron logs for more information. [ 578.227573] env[62183]: Removing descriptor: 16 [ 578.227573] env[62183]: ERROR nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ba424952-b1fa-476b-9694-d0e9b644f101, please check neutron logs for more information. [ 578.227573] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Traceback (most recent call last): [ 578.227573] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 578.227573] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] yield resources [ 578.227573] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.227573] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self.driver.spawn(context, instance, image_meta, [ 578.227573] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 578.227573] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.227573] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.227573] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] vm_ref = self.build_virtual_machine(instance, [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] for vif in network_info: [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] return self._sync_wrapper(fn, *args, **kwargs) [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self.wait() [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self[:] = self._gt.wait() [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] return self._exit_event.wait() [ 578.227951] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] result = hub.switch() [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] return self.greenlet.switch() [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] result = function(*args, **kwargs) [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] return func(*args, **kwargs) [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] raise e [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] nwinfo = self.network_api.allocate_for_instance( [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.228446] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] created_port_ids = self._update_ports_for_instance( [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] with excutils.save_and_reraise_exception(): [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self.force_reraise() [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] raise self.value [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] updated_port = self._update_port( [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] _ensure_no_port_binding_failure(port) [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.228834] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] raise exception.PortBindingFailed(port_id=port['id']) [ 578.229890] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] nova.exception.PortBindingFailed: Binding failed for port ba424952-b1fa-476b-9694-d0e9b644f101, please check neutron logs for more information. [ 578.229890] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] [ 578.229890] env[62183]: INFO nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Terminating instance [ 578.229890] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "refresh_cache-3800f34e-b3b6-4f8f-8df9-6f7266e7f901" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.229890] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquired lock "refresh_cache-3800f34e-b3b6-4f8f-8df9-6f7266e7f901" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.230137] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.308437] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.405521] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Successfully created port: a6802bb8-ad1d-4496-b8c6-277231254e26 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 578.759954] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.850073] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d72770b-23c2-49f0-93fd-12af71f8a8fb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.869403] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea3e19c-3110-4fd9-8dab-980389320440 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.902894] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2a4861-5f58-4784-8c2a-4fd2915b9a91 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.913147] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6224c858-dd55-4c3f-9585-d116631ec0ba {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.927620] env[62183]: DEBUG nova.compute.provider_tree [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.973742] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.236480] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Acquiring lock "1ad515a9-38ab-4094-9d71-c845dcb72cb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.236894] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Lock "1ad515a9-38ab-4094-9d71-c845dcb72cb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.290337] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "c2182423-e7fe-4ae2-be94-e88a1e49bab9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.290576] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "c2182423-e7fe-4ae2-be94-e88a1e49bab9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.326246] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.358919] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.359210] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.359367] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.359548] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.360060] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.360297] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.360558] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.360624] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.360822] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.360965] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.361175] env[62183]: DEBUG nova.virt.hardware [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.362660] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833c047b-985a-4a98-a8cb-6f8766b645a8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.373840] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c343d653-511d-41e4-9057-41a7060fb222 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.433491] env[62183]: DEBUG nova.scheduler.client.report [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 579.476128] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Releasing lock "refresh_cache-3800f34e-b3b6-4f8f-8df9-6f7266e7f901" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.476593] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 579.476813] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.477199] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9514dcb6-95f0-49bb-9715-32f85d840303 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.488256] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6116e2dd-cb12-4096-aa3a-94474642613a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.518732] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3800f34e-b3b6-4f8f-8df9-6f7266e7f901 could not be found. [ 579.518957] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.519156] env[62183]: INFO nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Took 0.04 seconds to destroy the instance on the hypervisor. [ 579.519448] env[62183]: DEBUG oslo.service.loopingcall [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.519696] env[62183]: DEBUG nova.compute.manager [-] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.519787] env[62183]: DEBUG nova.network.neutron [-] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.551623] env[62183]: DEBUG nova.network.neutron [-] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.686409] env[62183]: DEBUG nova.compute.manager [req-e6c58590-38e1-4633-b861-2d21a0b9e337 req-b43afcc7-5136-462c-a4af-526c983b7f99 service nova] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Received event network-changed-ba424952-b1fa-476b-9694-d0e9b644f101 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.686603] env[62183]: DEBUG nova.compute.manager [req-e6c58590-38e1-4633-b861-2d21a0b9e337 req-b43afcc7-5136-462c-a4af-526c983b7f99 service nova] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Refreshing instance network info cache due to event network-changed-ba424952-b1fa-476b-9694-d0e9b644f101. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.687214] env[62183]: DEBUG oslo_concurrency.lockutils [req-e6c58590-38e1-4633-b861-2d21a0b9e337 req-b43afcc7-5136-462c-a4af-526c983b7f99 service nova] Acquiring lock "refresh_cache-3800f34e-b3b6-4f8f-8df9-6f7266e7f901" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.687365] env[62183]: DEBUG oslo_concurrency.lockutils [req-e6c58590-38e1-4633-b861-2d21a0b9e337 req-b43afcc7-5136-462c-a4af-526c983b7f99 service nova] Acquired lock "refresh_cache-3800f34e-b3b6-4f8f-8df9-6f7266e7f901" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.687525] env[62183]: DEBUG nova.network.neutron [req-e6c58590-38e1-4633-b861-2d21a0b9e337 req-b43afcc7-5136-462c-a4af-526c983b7f99 service nova] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Refreshing network info cache for port ba424952-b1fa-476b-9694-d0e9b644f101 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 579.938838] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.646s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.939600] env[62183]: DEBUG nova.compute.manager [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 579.943324] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.691s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.944978] env[62183]: INFO nova.compute.claims [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 580.049139] env[62183]: ERROR nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a6802bb8-ad1d-4496-b8c6-277231254e26, please check neutron logs for more information. [ 580.049139] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 580.049139] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.049139] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 580.049139] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.049139] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 580.049139] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.049139] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 580.049139] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.049139] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 580.049139] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.049139] env[62183]: ERROR nova.compute.manager raise self.value [ 580.049139] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.049139] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 580.049139] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.049139] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 580.049782] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.049782] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 580.049782] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a6802bb8-ad1d-4496-b8c6-277231254e26, please check neutron logs for more information. [ 580.049782] env[62183]: ERROR nova.compute.manager [ 580.049782] env[62183]: Traceback (most recent call last): [ 580.049782] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 580.049782] env[62183]: listener.cb(fileno) [ 580.049782] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.049782] env[62183]: result = function(*args, **kwargs) [ 580.049782] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 580.049782] env[62183]: return func(*args, **kwargs) [ 580.049782] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.049782] env[62183]: raise e [ 580.049782] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.049782] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 580.049782] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.049782] env[62183]: created_port_ids = self._update_ports_for_instance( [ 580.049782] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.049782] env[62183]: with excutils.save_and_reraise_exception(): [ 580.049782] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.049782] env[62183]: self.force_reraise() [ 580.049782] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.049782] env[62183]: raise self.value [ 580.049782] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.049782] env[62183]: updated_port = self._update_port( [ 580.049782] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.049782] env[62183]: _ensure_no_port_binding_failure(port) [ 580.049782] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.049782] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 580.050917] env[62183]: nova.exception.PortBindingFailed: Binding failed for port a6802bb8-ad1d-4496-b8c6-277231254e26, please check neutron logs for more information. [ 580.050917] env[62183]: Removing descriptor: 14 [ 580.050917] env[62183]: ERROR nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a6802bb8-ad1d-4496-b8c6-277231254e26, please check neutron logs for more information. [ 580.050917] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Traceback (most recent call last): [ 580.050917] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 580.050917] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] yield resources [ 580.050917] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.050917] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self.driver.spawn(context, instance, image_meta, [ 580.050917] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 580.050917] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.050917] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.050917] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] vm_ref = self.build_virtual_machine(instance, [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] for vif in network_info: [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] return self._sync_wrapper(fn, *args, **kwargs) [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self.wait() [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self[:] = self._gt.wait() [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] return self._exit_event.wait() [ 580.051577] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] result = hub.switch() [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] return self.greenlet.switch() [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] result = function(*args, **kwargs) [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] return func(*args, **kwargs) [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] raise e [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] nwinfo = self.network_api.allocate_for_instance( [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 580.052138] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] created_port_ids = self._update_ports_for_instance( [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] with excutils.save_and_reraise_exception(): [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self.force_reraise() [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] raise self.value [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] updated_port = self._update_port( [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] _ensure_no_port_binding_failure(port) [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.052604] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] raise exception.PortBindingFailed(port_id=port['id']) [ 580.053846] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] nova.exception.PortBindingFailed: Binding failed for port a6802bb8-ad1d-4496-b8c6-277231254e26, please check neutron logs for more information. [ 580.053846] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] [ 580.053846] env[62183]: INFO nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Terminating instance [ 580.055249] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Acquiring lock "refresh_cache-2aa13e44-de7e-476a-aa3d-0675d38b5ef2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.055249] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Acquired lock "refresh_cache-2aa13e44-de7e-476a-aa3d-0675d38b5ef2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.055249] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 580.059085] env[62183]: DEBUG nova.network.neutron [-] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.358975] env[62183]: DEBUG nova.network.neutron [req-e6c58590-38e1-4633-b861-2d21a0b9e337 req-b43afcc7-5136-462c-a4af-526c983b7f99 service nova] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.453471] env[62183]: DEBUG nova.compute.utils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.456370] env[62183]: DEBUG nova.network.neutron [req-e6c58590-38e1-4633-b861-2d21a0b9e337 req-b43afcc7-5136-462c-a4af-526c983b7f99 service nova] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.457621] env[62183]: DEBUG nova.compute.manager [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Not allocating networking since 'none' was specified. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 580.566816] env[62183]: INFO nova.compute.manager [-] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Took 1.05 seconds to deallocate network for instance. [ 580.569368] env[62183]: DEBUG nova.compute.claims [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.569537] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.578528] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.666024] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.959991] env[62183]: DEBUG oslo_concurrency.lockutils [req-e6c58590-38e1-4633-b861-2d21a0b9e337 req-b43afcc7-5136-462c-a4af-526c983b7f99 service nova] Releasing lock "refresh_cache-3800f34e-b3b6-4f8f-8df9-6f7266e7f901" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.959991] env[62183]: DEBUG nova.compute.manager [req-e6c58590-38e1-4633-b861-2d21a0b9e337 req-b43afcc7-5136-462c-a4af-526c983b7f99 service nova] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Received event network-vif-deleted-ba424952-b1fa-476b-9694-d0e9b644f101 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.960562] env[62183]: DEBUG nova.compute.manager [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.169201] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Releasing lock "refresh_cache-2aa13e44-de7e-476a-aa3d-0675d38b5ef2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.169617] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 581.169808] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 581.170167] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f137944-d1cf-47fd-a684-8126a97642db {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.181144] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d365326-1573-435a-916d-07faaa380934 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.204920] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2aa13e44-de7e-476a-aa3d-0675d38b5ef2 could not be found. [ 581.205179] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 581.205366] env[62183]: INFO nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 581.205625] env[62183]: DEBUG oslo.service.loopingcall [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.208369] env[62183]: DEBUG nova.compute.manager [-] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.208466] env[62183]: DEBUG nova.network.neutron [-] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 581.239689] env[62183]: DEBUG nova.network.neutron [-] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.472287] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df81b8a-4705-43f9-9355-756c974ca92e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.480286] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd1b6267-777c-4aaa-a46b-20adfd7b2ba3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.515900] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6604320-0b9d-41c9-bca6-dc704e6e0745 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.524060] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde0433d-d9b6-4296-b618-e8c28b294fd2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.539040] env[62183]: DEBUG nova.compute.provider_tree [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.730021] env[62183]: DEBUG nova.compute.manager [req-df5968c9-c4dd-4a49-8b80-6aa1c99f7f9d req-1bb615a8-f5cc-441c-be3f-d7ce8a041f7e service nova] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Received event network-changed-a6802bb8-ad1d-4496-b8c6-277231254e26 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 581.730021] env[62183]: DEBUG nova.compute.manager [req-df5968c9-c4dd-4a49-8b80-6aa1c99f7f9d req-1bb615a8-f5cc-441c-be3f-d7ce8a041f7e service nova] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Refreshing instance network info cache due to event network-changed-a6802bb8-ad1d-4496-b8c6-277231254e26. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 581.730021] env[62183]: DEBUG oslo_concurrency.lockutils [req-df5968c9-c4dd-4a49-8b80-6aa1c99f7f9d req-1bb615a8-f5cc-441c-be3f-d7ce8a041f7e service nova] Acquiring lock "refresh_cache-2aa13e44-de7e-476a-aa3d-0675d38b5ef2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.730021] env[62183]: DEBUG oslo_concurrency.lockutils [req-df5968c9-c4dd-4a49-8b80-6aa1c99f7f9d req-1bb615a8-f5cc-441c-be3f-d7ce8a041f7e service nova] Acquired lock "refresh_cache-2aa13e44-de7e-476a-aa3d-0675d38b5ef2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.730021] env[62183]: DEBUG nova.network.neutron [req-df5968c9-c4dd-4a49-8b80-6aa1c99f7f9d req-1bb615a8-f5cc-441c-be3f-d7ce8a041f7e service nova] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Refreshing network info cache for port a6802bb8-ad1d-4496-b8c6-277231254e26 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 581.743485] env[62183]: DEBUG nova.network.neutron [-] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.977123] env[62183]: DEBUG nova.compute.manager [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.004662] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.004913] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.005087] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.005289] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.005461] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.005608] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.005877] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.005970] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.010049] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.010049] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.010049] env[62183]: DEBUG nova.virt.hardware [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.010049] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54cedae0-8ce1-44e1-84f3-6bc6dbd59bef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.016977] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f707e1b3-e4de-4d9f-adaa-125ce8d639c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.035168] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 582.043362] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 582.044270] env[62183]: DEBUG nova.scheduler.client.report [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 582.047203] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb651d96-bda7-4723-86cb-3f6172dee280 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.060978] env[62183]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 582.062020] env[62183]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62183) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 582.062020] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 582.062020] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Creating folder: Project (49f06b0868224065bcc5abd409bb73d4). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 582.062020] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f5c6f44f-6837-4fa0-9ebc-119696f71d3b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.073808] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Created folder: Project (49f06b0868224065bcc5abd409bb73d4) in parent group-v294392. [ 582.073808] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Creating folder: Instances. Parent ref: group-v294397. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 582.073808] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6b421e4-3955-4d83-a1d4-d6dc467ed2d8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.085855] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Created folder: Instances in parent group-v294397. [ 582.086124] env[62183]: DEBUG oslo.service.loopingcall [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.086328] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 582.086534] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3eb0c86-9104-4af6-982b-6d55d63066ba {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.105528] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 582.105528] env[62183]: value = "task-1386862" [ 582.105528] env[62183]: _type = "Task" [ 582.105528] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.114230] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386862, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.245842] env[62183]: INFO nova.compute.manager [-] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Took 1.04 seconds to deallocate network for instance. [ 582.248592] env[62183]: DEBUG nova.compute.claims [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 582.248765] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.254131] env[62183]: DEBUG nova.network.neutron [req-df5968c9-c4dd-4a49-8b80-6aa1c99f7f9d req-1bb615a8-f5cc-441c-be3f-d7ce8a041f7e service nova] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.418755] env[62183]: DEBUG nova.network.neutron [req-df5968c9-c4dd-4a49-8b80-6aa1c99f7f9d req-1bb615a8-f5cc-441c-be3f-d7ce8a041f7e service nova] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.552502] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.553625] env[62183]: DEBUG nova.compute.manager [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.557743] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.242s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.559230] env[62183]: INFO nova.compute.claims [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.618274] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386862, 'name': CreateVM_Task, 'duration_secs': 0.269107} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.619319] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 582.620593] env[62183]: DEBUG oslo_vmware.service [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5b1df9-8836-49da-8749-d7e65f445c46 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.630231] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.630231] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.630231] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 582.630231] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b426619-eba2-4ae7-99c6-d25fb677ad78 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.635220] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 582.635220] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52083e29-2a05-1179-0bab-50745bc6815e" [ 582.635220] env[62183]: _type = "Task" [ 582.635220] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.645508] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52083e29-2a05-1179-0bab-50745bc6815e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.921960] env[62183]: DEBUG oslo_concurrency.lockutils [req-df5968c9-c4dd-4a49-8b80-6aa1c99f7f9d req-1bb615a8-f5cc-441c-be3f-d7ce8a041f7e service nova] Releasing lock "refresh_cache-2aa13e44-de7e-476a-aa3d-0675d38b5ef2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.923128] env[62183]: DEBUG nova.compute.manager [req-df5968c9-c4dd-4a49-8b80-6aa1c99f7f9d req-1bb615a8-f5cc-441c-be3f-d7ce8a041f7e service nova] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Received event network-vif-deleted-a6802bb8-ad1d-4496-b8c6-277231254e26 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.069049] env[62183]: DEBUG nova.compute.utils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 583.071556] env[62183]: DEBUG nova.compute.manager [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Not allocating networking since 'none' was specified. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 583.148104] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.148371] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 583.148600] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.148743] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.149190] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 583.149754] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16c90246-7356-4c12-b75d-88478a686e20 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.574054] env[62183]: DEBUG nova.compute.manager [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.929266] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 583.929469] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 583.930347] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dea8659-7901-4b98-a9d7-86f58334c1fe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.945054] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d959a568-47e5-4f84-8063-558f05d2d915 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.950806] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 583.950806] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52538199-4653-d4ad-6b4b-6e82e3ecc2fc" [ 583.950806] env[62183]: _type = "Task" [ 583.950806] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.959421] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52538199-4653-d4ad-6b4b-6e82e3ecc2fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.969370] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Acquiring lock "541531de-d96c-47b5-a980-34dfc3e904aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.969605] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Lock "541531de-d96c-47b5-a980-34dfc3e904aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.013679] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e582c512-565a-4a25-8749-ca8e0fe079a1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.021676] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb41c1e-c070-4284-a578-a23a328baeb1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.050698] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee21d61-c7b2-4a06-9ee2-d5ca82de6d1d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.057722] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd93dd2-34ee-443c-89bd-74b1c98a7f42 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.071929] env[62183]: DEBUG nova.compute.provider_tree [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.462470] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52538199-4653-d4ad-6b4b-6e82e3ecc2fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.574219] env[62183]: DEBUG nova.scheduler.client.report [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 584.588764] env[62183]: DEBUG nova.compute.manager [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.612870] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.613135] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.613296] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.613476] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.613620] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.613764] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.613965] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.614135] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.614301] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.614458] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.614624] env[62183]: DEBUG nova.virt.hardware [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.615472] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd69a15-952e-44da-ad7d-bf38dbb3cdf0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.623347] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-250b5c88-645f-4395-934c-c8536d0ecccc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.636430] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 584.641728] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Creating folder: Project (2dfad2f57b1540c898ae9ec3a0832cd4). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 584.642096] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d4dfdfc-af6b-4388-99c1-5730e5501c8b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.650364] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Created folder: Project (2dfad2f57b1540c898ae9ec3a0832cd4) in parent group-v294392. [ 584.650532] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Creating folder: Instances. Parent ref: group-v294400. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 584.650729] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba05a33b-4499-4299-bfa0-940b754da5e4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.658342] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Created folder: Instances in parent group-v294400. [ 584.658552] env[62183]: DEBUG oslo.service.loopingcall [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.658724] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 584.658901] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3355e8da-2e1f-4ec8-ab1b-209650296cbb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.673929] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 584.673929] env[62183]: value = "task-1386865" [ 584.673929] env[62183]: _type = "Task" [ 584.673929] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.680665] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386865, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.962015] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Preparing fetch location {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 584.962301] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Creating directory with path [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 584.962533] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12c53376-e4b6-4d8b-a4ec-222fd47370b1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.079517] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.081103] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.083078] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.388s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.084500] env[62183]: INFO nova.compute.claims [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.191964] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386865, 'name': CreateVM_Task, 'duration_secs': 0.256453} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.192228] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 585.192569] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.192727] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.193116] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 585.193605] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-594d58de-1397-4c50-af3e-03266d76e949 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.198245] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 585.198245] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52330a31-3765-f039-f5d3-eef9a77953f2" [ 585.198245] env[62183]: _type = "Task" [ 585.198245] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.207717] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52330a31-3765-f039-f5d3-eef9a77953f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.466771] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Created directory with path [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 585.466999] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Fetch image to [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 585.467196] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Downloading image file data b250b36d-7961-4edf-b9ae-4be16b3e5ad2 to [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk on the data store datastore1 {{(pid=62183) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 585.468027] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fab5af-10c1-42aa-b534-ddc69b6a6da4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.476105] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8a2090-637a-41be-9ba0-60349d652464 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.485358] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0e8681-282f-4376-b5fa-d14bfc8ce5f6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.516429] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03f5287-ed3e-407c-b560-0da699626482 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.522315] env[62183]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-c780d472-5a55-4ad7-a4d4-1b00d7fc73c5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.545452] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Downloading image file data b250b36d-7961-4edf-b9ae-4be16b3e5ad2 to the data store datastore1 {{(pid=62183) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 585.589262] env[62183]: DEBUG nova.compute.utils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.593219] env[62183]: DEBUG oslo_vmware.rw_handles [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62183) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 585.594369] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.594550] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 585.662174] env[62183]: DEBUG nova.policy [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '245e9ab4b1924e94ab8c85a4a1e9695b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11a6dd7326ae4e64972f9de90792906b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.712073] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.712073] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 585.712340] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.035713] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Successfully created port: bc18163b-b86f-4c54-9b0e-1f68a35f444e {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.094036] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.251502] env[62183]: DEBUG oslo_vmware.rw_handles [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Completed reading data from the image iterator. {{(pid=62183) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 586.251502] env[62183]: DEBUG oslo_vmware.rw_handles [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 586.391778] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Downloaded image file data b250b36d-7961-4edf-b9ae-4be16b3e5ad2 to vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk on the data store datastore1 {{(pid=62183) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 586.395978] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Caching image {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 586.396272] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Copying Virtual Disk [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk to [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 586.396600] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c00dec60-5800-4751-a994-d5e105d8cbc0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.405234] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 586.405234] env[62183]: value = "task-1386866" [ 586.405234] env[62183]: _type = "Task" [ 586.405234] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.413809] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386866, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.543113] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d26dd83-8293-4014-a126-708a9b2f8063 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.552496] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa195d10-4cbf-481b-be49-a367654cc7f1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.581775] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8ffed7-ff9c-42ae-9c60-60e985f52c78 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.589392] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ac5844-8591-4c98-825f-adf0c87bc735 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.615865] env[62183]: DEBUG nova.compute.provider_tree [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.918738] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386866, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.932301] env[62183]: DEBUG nova.compute.manager [req-ba4c1df5-8eac-4117-9a87-bd44c3c51ce6 req-50ef6e7f-4961-4506-ac7d-d823cb225e42 service nova] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Received event network-changed-bc18163b-b86f-4c54-9b0e-1f68a35f444e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.932301] env[62183]: DEBUG nova.compute.manager [req-ba4c1df5-8eac-4117-9a87-bd44c3c51ce6 req-50ef6e7f-4961-4506-ac7d-d823cb225e42 service nova] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Refreshing instance network info cache due to event network-changed-bc18163b-b86f-4c54-9b0e-1f68a35f444e. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.932301] env[62183]: DEBUG oslo_concurrency.lockutils [req-ba4c1df5-8eac-4117-9a87-bd44c3c51ce6 req-50ef6e7f-4961-4506-ac7d-d823cb225e42 service nova] Acquiring lock "refresh_cache-4c9534c7-5498-46a1-b6c2-c93857031078" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.932591] env[62183]: DEBUG oslo_concurrency.lockutils [req-ba4c1df5-8eac-4117-9a87-bd44c3c51ce6 req-50ef6e7f-4961-4506-ac7d-d823cb225e42 service nova] Acquired lock "refresh_cache-4c9534c7-5498-46a1-b6c2-c93857031078" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.932787] env[62183]: DEBUG nova.network.neutron [req-ba4c1df5-8eac-4117-9a87-bd44c3c51ce6 req-50ef6e7f-4961-4506-ac7d-d823cb225e42 service nova] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Refreshing network info cache for port bc18163b-b86f-4c54-9b0e-1f68a35f444e {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 587.119264] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.121334] env[62183]: DEBUG nova.scheduler.client.report [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 587.129830] env[62183]: ERROR nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bc18163b-b86f-4c54-9b0e-1f68a35f444e, please check neutron logs for more information. [ 587.129830] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.129830] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.129830] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.129830] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.129830] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.129830] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.129830] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.129830] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.129830] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 587.129830] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.129830] env[62183]: ERROR nova.compute.manager raise self.value [ 587.129830] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.129830] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.129830] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.129830] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.131057] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.131057] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.131057] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bc18163b-b86f-4c54-9b0e-1f68a35f444e, please check neutron logs for more information. [ 587.131057] env[62183]: ERROR nova.compute.manager [ 587.131057] env[62183]: Traceback (most recent call last): [ 587.131057] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.131057] env[62183]: listener.cb(fileno) [ 587.131057] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.131057] env[62183]: result = function(*args, **kwargs) [ 587.131057] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.131057] env[62183]: return func(*args, **kwargs) [ 587.131057] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.131057] env[62183]: raise e [ 587.131057] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.131057] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 587.131057] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.131057] env[62183]: created_port_ids = self._update_ports_for_instance( [ 587.131057] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.131057] env[62183]: with excutils.save_and_reraise_exception(): [ 587.131057] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.131057] env[62183]: self.force_reraise() [ 587.131057] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.131057] env[62183]: raise self.value [ 587.131057] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.131057] env[62183]: updated_port = self._update_port( [ 587.131057] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.131057] env[62183]: _ensure_no_port_binding_failure(port) [ 587.131057] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.131057] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.132301] env[62183]: nova.exception.PortBindingFailed: Binding failed for port bc18163b-b86f-4c54-9b0e-1f68a35f444e, please check neutron logs for more information. [ 587.132301] env[62183]: Removing descriptor: 14 [ 587.147092] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.147439] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.147630] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.147830] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.148054] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.148260] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.148494] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.148871] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.148964] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.149125] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.149305] env[62183]: DEBUG nova.virt.hardware [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.150499] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051ea0e7-1892-46ee-9edd-4e836bd35521 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.160507] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0402317c-100f-4688-9e51-d42ab69eb296 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.174795] env[62183]: ERROR nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bc18163b-b86f-4c54-9b0e-1f68a35f444e, please check neutron logs for more information. [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Traceback (most recent call last): [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] yield resources [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self.driver.spawn(context, instance, image_meta, [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] vm_ref = self.build_virtual_machine(instance, [ 587.174795] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] for vif in network_info: [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] return self._sync_wrapper(fn, *args, **kwargs) [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self.wait() [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self[:] = self._gt.wait() [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] return self._exit_event.wait() [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 587.175217] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] current.throw(*self._exc) [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] result = function(*args, **kwargs) [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] return func(*args, **kwargs) [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] raise e [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] nwinfo = self.network_api.allocate_for_instance( [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] created_port_ids = self._update_ports_for_instance( [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] with excutils.save_and_reraise_exception(): [ 587.175684] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self.force_reraise() [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] raise self.value [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] updated_port = self._update_port( [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] _ensure_no_port_binding_failure(port) [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] raise exception.PortBindingFailed(port_id=port['id']) [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] nova.exception.PortBindingFailed: Binding failed for port bc18163b-b86f-4c54-9b0e-1f68a35f444e, please check neutron logs for more information. [ 587.176148] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] [ 587.176148] env[62183]: INFO nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Terminating instance [ 587.177749] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Acquiring lock "refresh_cache-4c9534c7-5498-46a1-b6c2-c93857031078" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.415781] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386866, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.996096} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.416084] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Copied Virtual Disk [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk to [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 587.416279] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Deleting the datastore file [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 587.416529] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ec2ff62-2ea8-48dd-8023-70e31ee2f724 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.423008] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 587.423008] env[62183]: value = "task-1386867" [ 587.423008] env[62183]: _type = "Task" [ 587.423008] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.430494] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.452221] env[62183]: DEBUG nova.network.neutron [req-ba4c1df5-8eac-4117-9a87-bd44c3c51ce6 req-50ef6e7f-4961-4506-ac7d-d823cb225e42 service nova] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.545633] env[62183]: DEBUG nova.network.neutron [req-ba4c1df5-8eac-4117-9a87-bd44c3c51ce6 req-50ef6e7f-4961-4506-ac7d-d823cb225e42 service nova] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.628325] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.628841] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 587.633027] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.272s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.933146] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.04038} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.933403] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 587.933606] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Moving file from [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880/b250b36d-7961-4edf-b9ae-4be16b3e5ad2 to [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2. {{(pid=62183) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 587.933976] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-9dee41d1-d901-4e42-a4ed-e0af7b8227f6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.942040] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 587.942040] env[62183]: value = "task-1386868" [ 587.942040] env[62183]: _type = "Task" [ 587.942040] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.949665] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386868, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.047999] env[62183]: DEBUG oslo_concurrency.lockutils [req-ba4c1df5-8eac-4117-9a87-bd44c3c51ce6 req-50ef6e7f-4961-4506-ac7d-d823cb225e42 service nova] Releasing lock "refresh_cache-4c9534c7-5498-46a1-b6c2-c93857031078" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.048453] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Acquired lock "refresh_cache-4c9534c7-5498-46a1-b6c2-c93857031078" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.048640] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.137301] env[62183]: DEBUG nova.compute.utils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.141637] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.141805] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 588.191065] env[62183]: DEBUG nova.policy [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e0eee67c7954c80b4e1ea43b0abbcb5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebcc716f958942b588a6bfde78d2c00d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 588.453933] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386868, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.047913} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.454250] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] File moved {{(pid=62183) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 588.454384] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Cleaning up location [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 588.454550] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Deleting the datastore file [datastore1] vmware_temp/3b46d131-4c28-4bfd-ab73-cfd163c7a880 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 588.454794] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37f95317-9050-4667-a217-5bb6e1453303 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.462562] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 588.462562] env[62183]: value = "task-1386869" [ 588.462562] env[62183]: _type = "Task" [ 588.462562] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.470181] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386869, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.508552] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Successfully created port: 62394ec9-e861-48dd-bca0-75b71c5f97f0 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.541118] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4b8428-4166-46d6-8d51-0fc54f2cf6e5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.548291] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8783dd61-9d58-4b48-a6c4-afc3987cf5a5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.579041] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.581248] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7576d0e-877c-4a05-a775-87f98d765b12 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.588216] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8714f5-14d7-4ed5-b7b1-1de5738b507c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.609647] env[62183]: DEBUG nova.compute.provider_tree [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.641898] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 588.725893] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.962834] env[62183]: DEBUG nova.compute.manager [req-249c2155-ff9d-482b-a8be-3f01882c7082 req-78e1b0e3-f9fd-44f5-9481-e8e3703f255e service nova] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Received event network-vif-deleted-bc18163b-b86f-4c54-9b0e-1f68a35f444e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.975365] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386869, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025034} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.975617] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 588.976346] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-430a34fa-7fc0-4c85-86c6-f9889dfae7e2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.982147] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 588.982147] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]527cf9e5-57a3-b4ee-f2dd-c9761f4f03c5" [ 588.982147] env[62183]: _type = "Task" [ 588.982147] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.989875] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527cf9e5-57a3-b4ee-f2dd-c9761f4f03c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.114027] env[62183]: DEBUG nova.scheduler.client.report [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 589.230460] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Releasing lock "refresh_cache-4c9534c7-5498-46a1-b6c2-c93857031078" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.230996] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.231302] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.231679] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2051135-646e-4659-ad5a-0174725f4535 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.243404] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ce14db-1b55-4722-a02b-117cedbd2848 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.269288] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c9534c7-5498-46a1-b6c2-c93857031078 could not be found. [ 589.269513] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 589.269694] env[62183]: INFO nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Took 0.04 seconds to destroy the instance on the hypervisor. [ 589.270338] env[62183]: DEBUG oslo.service.loopingcall [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.270338] env[62183]: DEBUG nova.compute.manager [-] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.270338] env[62183]: DEBUG nova.network.neutron [-] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.288333] env[62183]: DEBUG nova.network.neutron [-] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.493076] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527cf9e5-57a3-b4ee-f2dd-c9761f4f03c5, 'name': SearchDatastore_Task, 'duration_secs': 0.027338} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.493352] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.493594] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] a11cfd43-7516-4f98-8484-9ba2a0c5276d/a11cfd43-7516-4f98-8484-9ba2a0c5276d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 589.493862] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.494056] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 589.494262] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7acf5d37-6023-44a8-a25c-13282d515f06 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.496322] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e7b365c-4c18-449d-bb43-511ac36b56ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.503060] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 589.503060] env[62183]: value = "task-1386870" [ 589.503060] env[62183]: _type = "Task" [ 589.503060] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.508401] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 589.508587] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 589.509552] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95a19d20-4efd-478b-bdaf-9b84781cdde3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.514462] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386870, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.517367] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 589.517367] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]526c2fec-133c-1d9f-a1c2-79059661f583" [ 589.517367] env[62183]: _type = "Task" [ 589.517367] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.525074] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526c2fec-133c-1d9f-a1c2-79059661f583, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.612341] env[62183]: ERROR nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 62394ec9-e861-48dd-bca0-75b71c5f97f0, please check neutron logs for more information. [ 589.612341] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.612341] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.612341] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.612341] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.612341] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.612341] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.612341] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.612341] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.612341] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 589.612341] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.612341] env[62183]: ERROR nova.compute.manager raise self.value [ 589.612341] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.612341] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.612341] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.612341] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.613060] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.613060] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.613060] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 62394ec9-e861-48dd-bca0-75b71c5f97f0, please check neutron logs for more information. [ 589.613060] env[62183]: ERROR nova.compute.manager [ 589.613060] env[62183]: Traceback (most recent call last): [ 589.613060] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.613060] env[62183]: listener.cb(fileno) [ 589.613060] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.613060] env[62183]: result = function(*args, **kwargs) [ 589.613060] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.613060] env[62183]: return func(*args, **kwargs) [ 589.613060] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.613060] env[62183]: raise e [ 589.613060] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.613060] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 589.613060] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.613060] env[62183]: created_port_ids = self._update_ports_for_instance( [ 589.613060] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.613060] env[62183]: with excutils.save_and_reraise_exception(): [ 589.613060] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.613060] env[62183]: self.force_reraise() [ 589.613060] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.613060] env[62183]: raise self.value [ 589.613060] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.613060] env[62183]: updated_port = self._update_port( [ 589.613060] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.613060] env[62183]: _ensure_no_port_binding_failure(port) [ 589.613060] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.613060] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.614333] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 62394ec9-e861-48dd-bca0-75b71c5f97f0, please check neutron logs for more information. [ 589.614333] env[62183]: Removing descriptor: 14 [ 589.617550] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.618188] env[62183]: ERROR nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8456e319-d388-42ab-9fb8-731185839b57, please check neutron logs for more information. [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Traceback (most recent call last): [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self.driver.spawn(context, instance, image_meta, [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] vm_ref = self.build_virtual_machine(instance, [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.618188] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] for vif in network_info: [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] return self._sync_wrapper(fn, *args, **kwargs) [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self.wait() [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self[:] = self._gt.wait() [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] return self._exit_event.wait() [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] result = hub.switch() [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.618565] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] return self.greenlet.switch() [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] result = function(*args, **kwargs) [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] return func(*args, **kwargs) [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] raise e [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] nwinfo = self.network_api.allocate_for_instance( [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] created_port_ids = self._update_ports_for_instance( [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] with excutils.save_and_reraise_exception(): [ 589.618982] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] self.force_reraise() [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] raise self.value [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] updated_port = self._update_port( [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] _ensure_no_port_binding_failure(port) [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] raise exception.PortBindingFailed(port_id=port['id']) [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] nova.exception.PortBindingFailed: Binding failed for port 8456e319-d388-42ab-9fb8-731185839b57, please check neutron logs for more information. [ 589.619418] env[62183]: ERROR nova.compute.manager [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] [ 589.619741] env[62183]: DEBUG nova.compute.utils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Binding failed for port 8456e319-d388-42ab-9fb8-731185839b57, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 589.620390] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.870s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.624012] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Build of instance 97f8b7df-50c6-43c6-ac0d-5b61541f750b was re-scheduled: Binding failed for port 8456e319-d388-42ab-9fb8-731185839b57, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 589.624441] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 589.624662] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Acquiring lock "refresh_cache-97f8b7df-50c6-43c6-ac0d-5b61541f750b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.624806] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Acquired lock "refresh_cache-97f8b7df-50c6-43c6-ac0d-5b61541f750b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.624962] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.661666] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 589.683831] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.683973] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.684170] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.684366] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.684514] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.684658] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.684864] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.685028] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.685200] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.685357] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.685523] env[62183]: DEBUG nova.virt.hardware [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.686375] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2346704b-0e24-4c81-984a-f498b9d7b357 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.694311] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48009226-5247-46db-a94d-f2caadebb8ff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.708040] env[62183]: ERROR nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 62394ec9-e861-48dd-bca0-75b71c5f97f0, please check neutron logs for more information. [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Traceback (most recent call last): [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] yield resources [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self.driver.spawn(context, instance, image_meta, [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] vm_ref = self.build_virtual_machine(instance, [ 589.708040] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] for vif in network_info: [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] return self._sync_wrapper(fn, *args, **kwargs) [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self.wait() [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self[:] = self._gt.wait() [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] return self._exit_event.wait() [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 589.708458] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] current.throw(*self._exc) [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] result = function(*args, **kwargs) [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] return func(*args, **kwargs) [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] raise e [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] nwinfo = self.network_api.allocate_for_instance( [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] created_port_ids = self._update_ports_for_instance( [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] with excutils.save_and_reraise_exception(): [ 589.708842] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self.force_reraise() [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] raise self.value [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] updated_port = self._update_port( [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] _ensure_no_port_binding_failure(port) [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] raise exception.PortBindingFailed(port_id=port['id']) [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] nova.exception.PortBindingFailed: Binding failed for port 62394ec9-e861-48dd-bca0-75b71c5f97f0, please check neutron logs for more information. [ 589.709274] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] [ 589.709274] env[62183]: INFO nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Terminating instance [ 589.710767] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.710930] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.711167] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.791014] env[62183]: DEBUG nova.network.neutron [-] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.012903] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386870, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.026305] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526c2fec-133c-1d9f-a1c2-79059661f583, 'name': SearchDatastore_Task, 'duration_secs': 0.017561} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.027522] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b13b1e77-7b23-4150-bb6e-906f9496fba3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.032247] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 590.032247] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52704249-e634-d3bf-1773-786197f7d3be" [ 590.032247] env[62183]: _type = "Task" [ 590.032247] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.040732] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52704249-e634-d3bf-1773-786197f7d3be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.146095] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.231818] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.275335] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.292703] env[62183]: INFO nova.compute.manager [-] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Took 1.02 seconds to deallocate network for instance. [ 590.298027] env[62183]: DEBUG nova.compute.claims [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.298027] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.350967] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.513488] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386870, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649412} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.513488] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] a11cfd43-7516-4f98-8484-9ba2a0c5276d/a11cfd43-7516-4f98-8484-9ba2a0c5276d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 590.513736] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 590.517020] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57b5f733-0fb2-4118-8c31-202d2c35067b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.519999] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 590.519999] env[62183]: value = "task-1386871" [ 590.519999] env[62183]: _type = "Task" [ 590.519999] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.533026] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386871, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.541288] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52704249-e634-d3bf-1773-786197f7d3be, 'name': SearchDatastore_Task, 'duration_secs': 0.069981} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.541703] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.541980] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 590.542361] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1a3bda2-816e-41c4-adae-45aa6024ec35 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.545630] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e1d479-ba95-4b95-b972-785de2f90752 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.552918] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e858c679-8ec9-4f4e-9b13-11ae6f5260e5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.555814] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 590.555814] env[62183]: value = "task-1386872" [ 590.555814] env[62183]: _type = "Task" [ 590.555814] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.583382] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9b55c1-2fc8-4988-bb3b-b1be4c2b6d91 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.591025] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386872, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.594652] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11efd74d-52e6-40d6-8794-8fe34c97a3e8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.607613] env[62183]: DEBUG nova.compute.provider_tree [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.777643] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Releasing lock "refresh_cache-97f8b7df-50c6-43c6-ac0d-5b61541f750b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.777955] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 590.778238] env[62183]: DEBUG nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.778433] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.791467] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.854301] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.854681] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.854728] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 590.855102] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7fea2ff9-b8d4-4236-92bc-2d56d9b1e2d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.864207] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc10a9b9-75b2-4560-a371-d333998de8c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.888119] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5 could not be found. [ 590.888330] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.888510] env[62183]: INFO nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 590.888742] env[62183]: DEBUG oslo.service.loopingcall [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.888953] env[62183]: DEBUG nova.compute.manager [-] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.889068] env[62183]: DEBUG nova.network.neutron [-] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.904873] env[62183]: DEBUG nova.network.neutron [-] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.994063] env[62183]: DEBUG nova.compute.manager [req-64c412ea-4089-423a-95cf-0a96fc533c36 req-ec13781b-77c2-46f3-ab8a-3694b775f79a service nova] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Received event network-changed-62394ec9-e861-48dd-bca0-75b71c5f97f0 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.994063] env[62183]: DEBUG nova.compute.manager [req-64c412ea-4089-423a-95cf-0a96fc533c36 req-ec13781b-77c2-46f3-ab8a-3694b775f79a service nova] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Refreshing instance network info cache due to event network-changed-62394ec9-e861-48dd-bca0-75b71c5f97f0. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 590.994063] env[62183]: DEBUG oslo_concurrency.lockutils [req-64c412ea-4089-423a-95cf-0a96fc533c36 req-ec13781b-77c2-46f3-ab8a-3694b775f79a service nova] Acquiring lock "refresh_cache-85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.994063] env[62183]: DEBUG oslo_concurrency.lockutils [req-64c412ea-4089-423a-95cf-0a96fc533c36 req-ec13781b-77c2-46f3-ab8a-3694b775f79a service nova] Acquired lock "refresh_cache-85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.994639] env[62183]: DEBUG nova.network.neutron [req-64c412ea-4089-423a-95cf-0a96fc533c36 req-ec13781b-77c2-46f3-ab8a-3694b775f79a service nova] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Refreshing network info cache for port 62394ec9-e861-48dd-bca0-75b71c5f97f0 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 591.031019] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386871, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08907} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.031196] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 591.032194] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377b1f30-4514-45e4-8efc-ef28164ac97a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.054830] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Reconfiguring VM instance instance-00000010 to attach disk [datastore1] a11cfd43-7516-4f98-8484-9ba2a0c5276d/a11cfd43-7516-4f98-8484-9ba2a0c5276d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 591.055122] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca400a32-0c76-43d9-86aa-4c357e418ee8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.077833] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386872, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.079399] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 591.079399] env[62183]: value = "task-1386873" [ 591.079399] env[62183]: _type = "Task" [ 591.079399] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.087773] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386873, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.111106] env[62183]: DEBUG nova.scheduler.client.report [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 591.293580] env[62183]: DEBUG nova.network.neutron [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.407165] env[62183]: DEBUG nova.network.neutron [-] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.522210] env[62183]: DEBUG nova.network.neutron [req-64c412ea-4089-423a-95cf-0a96fc533c36 req-ec13781b-77c2-46f3-ab8a-3694b775f79a service nova] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.579221] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386872, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689532} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.579488] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 591.579697] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 591.580018] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0019a072-852e-43b9-8775-8f90621b9c9b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.590058] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386873, 'name': ReconfigVM_Task, 'duration_secs': 0.297127} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.591386] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Reconfigured VM instance instance-00000010 to attach disk [datastore1] a11cfd43-7516-4f98-8484-9ba2a0c5276d/a11cfd43-7516-4f98-8484-9ba2a0c5276d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 591.592183] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 591.592183] env[62183]: value = "task-1386874" [ 591.592183] env[62183]: _type = "Task" [ 591.592183] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.592371] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e9519e5c-b49f-4008-9b54-0df9e878218a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.603380] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386874, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.604795] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 591.604795] env[62183]: value = "task-1386875" [ 591.604795] env[62183]: _type = "Task" [ 591.604795] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.608369] env[62183]: DEBUG nova.network.neutron [req-64c412ea-4089-423a-95cf-0a96fc533c36 req-ec13781b-77c2-46f3-ab8a-3694b775f79a service nova] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.614746] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386875, 'name': Rename_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.616522] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.617167] env[62183]: ERROR nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6481fc64-927e-40f0-959d-2643258b26b5, please check neutron logs for more information. [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Traceback (most recent call last): [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self.driver.spawn(context, instance, image_meta, [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] vm_ref = self.build_virtual_machine(instance, [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.617167] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] for vif in network_info: [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] return self._sync_wrapper(fn, *args, **kwargs) [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self.wait() [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self[:] = self._gt.wait() [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] return self._exit_event.wait() [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] result = hub.switch() [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.617548] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] return self.greenlet.switch() [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] result = function(*args, **kwargs) [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] return func(*args, **kwargs) [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] raise e [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] nwinfo = self.network_api.allocate_for_instance( [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] created_port_ids = self._update_ports_for_instance( [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] with excutils.save_and_reraise_exception(): [ 591.618047] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] self.force_reraise() [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] raise self.value [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] updated_port = self._update_port( [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] _ensure_no_port_binding_failure(port) [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] raise exception.PortBindingFailed(port_id=port['id']) [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] nova.exception.PortBindingFailed: Binding failed for port 6481fc64-927e-40f0-959d-2643258b26b5, please check neutron logs for more information. [ 591.618491] env[62183]: ERROR nova.compute.manager [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] [ 591.618851] env[62183]: DEBUG nova.compute.utils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Binding failed for port 6481fc64-927e-40f0-959d-2643258b26b5, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 591.619527] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.388s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.620981] env[62183]: INFO nova.compute.claims [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 591.623626] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Build of instance 5ee79852-b725-4df7-8a64-38ad1805bb66 was re-scheduled: Binding failed for port 6481fc64-927e-40f0-959d-2643258b26b5, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 591.624095] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 591.624323] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Acquiring lock "refresh_cache-5ee79852-b725-4df7-8a64-38ad1805bb66" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.624472] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Acquired lock "refresh_cache-5ee79852-b725-4df7-8a64-38ad1805bb66" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.624630] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.796322] env[62183]: INFO nova.compute.manager [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] [instance: 97f8b7df-50c6-43c6-ac0d-5b61541f750b] Took 1.02 seconds to deallocate network for instance. [ 591.911835] env[62183]: INFO nova.compute.manager [-] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Took 1.02 seconds to deallocate network for instance. [ 591.914161] env[62183]: DEBUG nova.compute.claims [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 591.914340] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.107073] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386874, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.134106} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.111414] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 592.111879] env[62183]: DEBUG oslo_concurrency.lockutils [req-64c412ea-4089-423a-95cf-0a96fc533c36 req-ec13781b-77c2-46f3-ab8a-3694b775f79a service nova] Releasing lock "refresh_cache-85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.112119] env[62183]: DEBUG nova.compute.manager [req-64c412ea-4089-423a-95cf-0a96fc533c36 req-ec13781b-77c2-46f3-ab8a-3694b775f79a service nova] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Received event network-vif-deleted-62394ec9-e861-48dd-bca0-75b71c5f97f0 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.112976] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac30fb7-eee3-42c0-ac54-9214e5138fdd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.120336] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386875, 'name': Rename_Task, 'duration_secs': 0.260992} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.127202] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 592.136032] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Reconfiguring VM instance instance-00000011 to attach disk [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 592.136032] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b9daeb1-64e9-4a97-ad1d-9fa405b8ce2b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.141510] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-797f2457-4dd0-4bf5-bbdf-9cda3583af6f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.161136] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 592.161136] env[62183]: value = "task-1386876" [ 592.161136] env[62183]: _type = "Task" [ 592.161136] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.166035] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 592.166035] env[62183]: value = "task-1386877" [ 592.166035] env[62183]: _type = "Task" [ 592.166035] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.172553] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.173356] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.181175] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.268191] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.675800] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386876, 'name': PowerOnVM_Task} progress is 98%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.678813] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386877, 'name': ReconfigVM_Task, 'duration_secs': 0.33591} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.679079] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Reconfigured VM instance instance-00000011 to attach disk [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 592.679692] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a307c1c-b95a-4b29-86ad-467fab8abfe0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.685792] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 592.685792] env[62183]: value = "task-1386878" [ 592.685792] env[62183]: _type = "Task" [ 592.685792] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.693931] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386878, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.772206] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Releasing lock "refresh_cache-5ee79852-b725-4df7-8a64-38ad1805bb66" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.772645] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 592.772910] env[62183]: DEBUG nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.773274] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.795036] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.830550] env[62183]: INFO nova.scheduler.client.report [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Deleted allocations for instance 97f8b7df-50c6-43c6-ac0d-5b61541f750b [ 593.176391] env[62183]: DEBUG oslo_vmware.api [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386876, 'name': PowerOnVM_Task, 'duration_secs': 0.847977} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.179392] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 593.179614] env[62183]: INFO nova.compute.manager [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Took 11.20 seconds to spawn the instance on the hypervisor. [ 593.179870] env[62183]: DEBUG nova.compute.manager [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 593.181278] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f4bef9-efd9-4e3f-88a6-9dbcad9f8a99 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.206275] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386878, 'name': Rename_Task, 'duration_secs': 0.132603} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.206544] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 593.206788] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ad07a74-1941-411f-a94a-856502c0bfe0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.213577] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 593.213577] env[62183]: value = "task-1386879" [ 593.213577] env[62183]: _type = "Task" [ 593.213577] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.224500] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386879, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.249301] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7086055f-1fc0-429d-995f-5b8af054394e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.257018] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41886a4-3c3e-4e28-97c5-30d5526557ab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.288697] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df0ea4b-dd4c-4710-832f-ed0d718c9cbf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.298337] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26eb6db8-2f51-4bf4-86e1-c352a9f11b0f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.302967] env[62183]: DEBUG nova.network.neutron [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.320235] env[62183]: DEBUG nova.compute.provider_tree [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.343791] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a2b90b23-b8db-4476-8a38-748ca97522d3 tempest-ServerRescueTestJSON-795009787 tempest-ServerRescueTestJSON-795009787-project-member] Lock "97f8b7df-50c6-43c6-ac0d-5b61541f750b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.463s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.711342] env[62183]: INFO nova.compute.manager [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Took 32.27 seconds to build instance. [ 593.726725] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386879, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.809410] env[62183]: INFO nova.compute.manager [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] [instance: 5ee79852-b725-4df7-8a64-38ad1805bb66] Took 1.04 seconds to deallocate network for instance. [ 593.823757] env[62183]: DEBUG nova.scheduler.client.report [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 593.850552] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 593.867418] env[62183]: DEBUG nova.compute.manager [None req-ea7f38fc-6eab-4890-b898-03b305203bc5 tempest-ServerDiagnosticsV248Test-2011752479 tempest-ServerDiagnosticsV248Test-2011752479-project-admin] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 593.869436] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4fcd34-7cd3-413e-a988-c2579f47b9ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.879511] env[62183]: INFO nova.compute.manager [None req-ea7f38fc-6eab-4890-b898-03b305203bc5 tempest-ServerDiagnosticsV248Test-2011752479 tempest-ServerDiagnosticsV248Test-2011752479-project-admin] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Retrieving diagnostics [ 593.881898] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92cb7802-885a-4e61-ab8c-38a0a4f1ee5c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.216411] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bcd5c302-39af-4fca-9aa4-03dc30d2f768 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "a11cfd43-7516-4f98-8484-9ba2a0c5276d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.880s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.227515] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386879, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.329102] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.709s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.329641] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 594.332145] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.071s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.384634] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.722585] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 594.728757] env[62183]: DEBUG oslo_vmware.api [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386879, 'name': PowerOnVM_Task, 'duration_secs': 1.14381} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.729280] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 594.729518] env[62183]: INFO nova.compute.manager [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Took 10.14 seconds to spawn the instance on the hypervisor. [ 594.729715] env[62183]: DEBUG nova.compute.manager [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 594.733020] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e49b9de-cac4-445d-93aa-9f1b8c28ad8f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.843211] env[62183]: DEBUG nova.compute.utils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 594.844690] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 594.844861] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 594.851244] env[62183]: INFO nova.scheduler.client.report [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Deleted allocations for instance 5ee79852-b725-4df7-8a64-38ad1805bb66 [ 594.955286] env[62183]: DEBUG nova.policy [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66807ca4e49d45bc886b8eba13757abe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbd19d0c9b1e4090bd9cafd5610ecd39', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.249156] env[62183]: INFO nova.compute.manager [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Took 32.01 seconds to build instance. [ 595.250920] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.326061] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41edc4ac-8c71-412f-b8d6-31cfb874cb51 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.331321] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Successfully created port: d4e93cda-fa04-4fb5-8636-85f18a5771c9 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 595.336400] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed52e369-a6ba-4d6e-a5fe-4e35350e2177 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.368211] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 595.370955] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c31a5f70-0b32-41ad-bd0d-ba1487b7f298 tempest-ServersWithSpecificFlavorTestJSON-1873997066 tempest-ServersWithSpecificFlavorTestJSON-1873997066-project-member] Lock "5ee79852-b725-4df7-8a64-38ad1805bb66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.398s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.375076] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1474d0fd-8ad6-42b6-9d3a-b1209b9e2e09 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.384927] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff011116-f508-4e54-a975-74aca9dee861 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.401696] env[62183]: DEBUG nova.compute.provider_tree [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.752793] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d3f5a8-e037-409e-abcd-25cd37a6f461 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "a43e24b0-c0f1-4e02-894d-0d2c8ed884d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.618s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.881060] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.908585] env[62183]: DEBUG nova.scheduler.client.report [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 596.160013] env[62183]: INFO nova.compute.manager [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Rebuilding instance [ 596.211595] env[62183]: DEBUG nova.compute.manager [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 596.212388] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656288ff-2813-422f-9a8d-15197fd4aff3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.255348] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.379124] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 596.383595] env[62183]: DEBUG nova.compute.manager [req-3445a889-0e7e-4b8f-935e-4422a26a0c5d req-915a7f38-0743-40ec-ae31-4808c799d00d service nova] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Received event network-changed-d4e93cda-fa04-4fb5-8636-85f18a5771c9 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.383775] env[62183]: DEBUG nova.compute.manager [req-3445a889-0e7e-4b8f-935e-4422a26a0c5d req-915a7f38-0743-40ec-ae31-4808c799d00d service nova] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Refreshing instance network info cache due to event network-changed-d4e93cda-fa04-4fb5-8636-85f18a5771c9. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 596.383984] env[62183]: DEBUG oslo_concurrency.lockutils [req-3445a889-0e7e-4b8f-935e-4422a26a0c5d req-915a7f38-0743-40ec-ae31-4808c799d00d service nova] Acquiring lock "refresh_cache-1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.384141] env[62183]: DEBUG oslo_concurrency.lockutils [req-3445a889-0e7e-4b8f-935e-4422a26a0c5d req-915a7f38-0743-40ec-ae31-4808c799d00d service nova] Acquired lock "refresh_cache-1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.384299] env[62183]: DEBUG nova.network.neutron [req-3445a889-0e7e-4b8f-935e-4422a26a0c5d req-915a7f38-0743-40ec-ae31-4808c799d00d service nova] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Refreshing network info cache for port d4e93cda-fa04-4fb5-8636-85f18a5771c9 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 596.403842] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.411313] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 596.411424] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 596.411580] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 596.412309] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 596.412309] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 596.415488] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 596.415488] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 596.415488] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 596.415707] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 596.415880] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 596.416113] env[62183]: DEBUG nova.virt.hardware [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.416688] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.085s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.417311] env[62183]: ERROR nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 98996e3e-45ac-4030-b4cb-d0490643c52c, please check neutron logs for more information. [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Traceback (most recent call last): [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self.driver.spawn(context, instance, image_meta, [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] vm_ref = self.build_virtual_machine(instance, [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.417311] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] for vif in network_info: [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] return self._sync_wrapper(fn, *args, **kwargs) [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self.wait() [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self[:] = self._gt.wait() [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] return self._exit_event.wait() [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] result = hub.switch() [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.417577] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] return self.greenlet.switch() [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] result = function(*args, **kwargs) [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] return func(*args, **kwargs) [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] raise e [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] nwinfo = self.network_api.allocate_for_instance( [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] created_port_ids = self._update_ports_for_instance( [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] with excutils.save_and_reraise_exception(): [ 596.417855] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] self.force_reraise() [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] raise self.value [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] updated_port = self._update_port( [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] _ensure_no_port_binding_failure(port) [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] raise exception.PortBindingFailed(port_id=port['id']) [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] nova.exception.PortBindingFailed: Binding failed for port 98996e3e-45ac-4030-b4cb-d0490643c52c, please check neutron logs for more information. [ 596.418156] env[62183]: ERROR nova.compute.manager [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] [ 596.418395] env[62183]: DEBUG nova.compute.utils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Binding failed for port 98996e3e-45ac-4030-b4cb-d0490643c52c, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 596.419781] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc164cab-b3b6-438b-96e8-844155e39a69 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.422942] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Build of instance 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff was re-scheduled: Binding failed for port 98996e3e-45ac-4030-b4cb-d0490643c52c, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 596.423019] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 596.423250] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.426352] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.426352] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.426352] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.855s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.435942] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24528862-caf2-40d0-924f-62e2a4c3d9fa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.444486] env[62183]: ERROR nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d4e93cda-fa04-4fb5-8636-85f18a5771c9, please check neutron logs for more information. [ 596.444486] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.444486] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.444486] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.444486] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.444486] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.444486] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.444486] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.444486] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.444486] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 596.444486] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.444486] env[62183]: ERROR nova.compute.manager raise self.value [ 596.444486] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.444486] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.444486] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.444486] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.444857] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.444857] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.444857] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d4e93cda-fa04-4fb5-8636-85f18a5771c9, please check neutron logs for more information. [ 596.444857] env[62183]: ERROR nova.compute.manager [ 596.444857] env[62183]: Traceback (most recent call last): [ 596.444857] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.444857] env[62183]: listener.cb(fileno) [ 596.444857] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.444857] env[62183]: result = function(*args, **kwargs) [ 596.444857] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.444857] env[62183]: return func(*args, **kwargs) [ 596.444857] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.444857] env[62183]: raise e [ 596.444857] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.444857] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 596.444857] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.444857] env[62183]: created_port_ids = self._update_ports_for_instance( [ 596.444857] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.444857] env[62183]: with excutils.save_and_reraise_exception(): [ 596.444857] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.444857] env[62183]: self.force_reraise() [ 596.444857] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.444857] env[62183]: raise self.value [ 596.444857] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.444857] env[62183]: updated_port = self._update_port( [ 596.444857] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.444857] env[62183]: _ensure_no_port_binding_failure(port) [ 596.444857] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.444857] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.449237] env[62183]: nova.exception.PortBindingFailed: Binding failed for port d4e93cda-fa04-4fb5-8636-85f18a5771c9, please check neutron logs for more information. [ 596.449237] env[62183]: Removing descriptor: 16 [ 596.454924] env[62183]: ERROR nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d4e93cda-fa04-4fb5-8636-85f18a5771c9, please check neutron logs for more information. [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Traceback (most recent call last): [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] yield resources [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self.driver.spawn(context, instance, image_meta, [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] vm_ref = self.build_virtual_machine(instance, [ 596.454924] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] for vif in network_info: [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] return self._sync_wrapper(fn, *args, **kwargs) [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self.wait() [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self[:] = self._gt.wait() [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] return self._exit_event.wait() [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 596.455270] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] current.throw(*self._exc) [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] result = function(*args, **kwargs) [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] return func(*args, **kwargs) [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] raise e [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] nwinfo = self.network_api.allocate_for_instance( [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] created_port_ids = self._update_ports_for_instance( [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] with excutils.save_and_reraise_exception(): [ 596.455586] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self.force_reraise() [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] raise self.value [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] updated_port = self._update_port( [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] _ensure_no_port_binding_failure(port) [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] raise exception.PortBindingFailed(port_id=port['id']) [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] nova.exception.PortBindingFailed: Binding failed for port d4e93cda-fa04-4fb5-8636-85f18a5771c9, please check neutron logs for more information. [ 596.455986] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] [ 596.455986] env[62183]: INFO nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Terminating instance [ 596.458478] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Acquiring lock "refresh_cache-1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.724272] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 596.724580] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef3cd9c2-d021-4882-8d9e-51d42c2813e1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.735020] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 596.735020] env[62183]: value = "task-1386880" [ 596.735020] env[62183]: _type = "Task" [ 596.735020] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.743606] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.781781] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.915649] env[62183]: DEBUG nova.network.neutron [req-3445a889-0e7e-4b8f-935e-4422a26a0c5d req-915a7f38-0743-40ec-ae31-4808c799d00d service nova] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.954970] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.093831] env[62183]: DEBUG nova.network.neutron [req-3445a889-0e7e-4b8f-935e-4422a26a0c5d req-915a7f38-0743-40ec-ae31-4808c799d00d service nova] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.129493] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.245977] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386880, 'name': PowerOffVM_Task, 'duration_secs': 0.130812} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.246151] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 597.246391] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 597.247249] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a31426-05e8-4533-a694-883e372887e4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.256573] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 597.256870] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3f0f4ae-a638-4409-80f2-3a6e87f069bc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.281506] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 597.281506] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 597.281506] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Deleting the datastore file [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 597.281506] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccc31888-032f-4c1a-b97b-c13fb340b996 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.289620] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 597.289620] env[62183]: value = "task-1386882" [ 597.289620] env[62183]: _type = "Task" [ 597.289620] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.299981] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.371977] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99273b07-61d0-45ca-9774-82c16e98edbb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.382873] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38841bfc-732e-4f2d-8294-020837cb3aa8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.415280] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e4e23b-e560-42e1-b11a-ad681939bc1d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.423048] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873686c6-9baf-4071-afe8-49c433359f73 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.439095] env[62183]: DEBUG nova.compute.provider_tree [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.597177] env[62183]: DEBUG oslo_concurrency.lockutils [req-3445a889-0e7e-4b8f-935e-4422a26a0c5d req-915a7f38-0743-40ec-ae31-4808c799d00d service nova] Releasing lock "refresh_cache-1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.597650] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Acquired lock "refresh_cache-1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.600070] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.632183] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.632415] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 597.632598] env[62183]: DEBUG nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.632756] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 597.646014] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.800171] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140827} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.800453] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 597.800632] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 597.800797] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 597.943480] env[62183]: DEBUG nova.scheduler.client.report [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 598.119453] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.148835] env[62183]: DEBUG nova.network.neutron [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.230214] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.413078] env[62183]: DEBUG nova.compute.manager [req-ec0db241-8fcd-45b2-b48b-25818c493fa7 req-a951cbef-d1b7-4a22-89b2-af61e39413e3 service nova] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Received event network-vif-deleted-d4e93cda-fa04-4fb5-8636-85f18a5771c9 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.452751] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.453381] env[62183]: ERROR nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ba424952-b1fa-476b-9694-d0e9b644f101, please check neutron logs for more information. [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Traceback (most recent call last): [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self.driver.spawn(context, instance, image_meta, [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] vm_ref = self.build_virtual_machine(instance, [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.453381] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] for vif in network_info: [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] return self._sync_wrapper(fn, *args, **kwargs) [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self.wait() [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self[:] = self._gt.wait() [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] return self._exit_event.wait() [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] result = hub.switch() [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.453663] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] return self.greenlet.switch() [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] result = function(*args, **kwargs) [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] return func(*args, **kwargs) [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] raise e [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] nwinfo = self.network_api.allocate_for_instance( [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] created_port_ids = self._update_ports_for_instance( [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] with excutils.save_and_reraise_exception(): [ 598.453925] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] self.force_reraise() [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] raise self.value [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] updated_port = self._update_port( [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] _ensure_no_port_binding_failure(port) [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] raise exception.PortBindingFailed(port_id=port['id']) [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] nova.exception.PortBindingFailed: Binding failed for port ba424952-b1fa-476b-9694-d0e9b644f101, please check neutron logs for more information. [ 598.454328] env[62183]: ERROR nova.compute.manager [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] [ 598.454587] env[62183]: DEBUG nova.compute.utils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Binding failed for port ba424952-b1fa-476b-9694-d0e9b644f101, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 598.455345] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.206s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.458557] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Build of instance 3800f34e-b3b6-4f8f-8df9-6f7266e7f901 was re-scheduled: Binding failed for port ba424952-b1fa-476b-9694-d0e9b644f101, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 598.458695] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 598.458838] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "refresh_cache-3800f34e-b3b6-4f8f-8df9-6f7266e7f901" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.458992] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquired lock "refresh_cache-3800f34e-b3b6-4f8f-8df9-6f7266e7f901" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.459196] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.652342] env[62183]: INFO nova.compute.manager [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff] Took 1.02 seconds to deallocate network for instance. [ 598.732571] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Releasing lock "refresh_cache-1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.733021] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.733221] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 598.733521] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff4448ba-b8a6-4b03-b0ea-1d45c4e40618 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.743781] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2231af-14f6-42d4-a0f0-1bd1717a4996 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.768513] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe could not be found. [ 598.768774] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 598.768992] env[62183]: INFO nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Took 0.04 seconds to destroy the instance on the hypervisor. [ 598.769287] env[62183]: DEBUG oslo.service.loopingcall [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.769548] env[62183]: DEBUG nova.compute.manager [-] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.769678] env[62183]: DEBUG nova.network.neutron [-] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.785412] env[62183]: DEBUG nova.network.neutron [-] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.834645] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.834645] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.834963] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.834963] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.835104] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.835302] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.835556] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.835720] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.835973] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.836163] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.836349] env[62183]: DEBUG nova.virt.hardware [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.837226] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8d422b-038a-4d78-a188-1440d57b84ea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.845308] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4ef969-c42b-461c-9841-9951ae5feca9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.858976] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 598.864422] env[62183]: DEBUG oslo.service.loopingcall [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.865091] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 598.865487] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58f0782e-6380-4c47-a4d8-91e101a6cb9c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.881375] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 598.881375] env[62183]: value = "task-1386883" [ 598.881375] env[62183]: _type = "Task" [ 598.881375] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.888776] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386883, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.983335] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.078416] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.287886] env[62183]: DEBUG nova.network.neutron [-] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.393139] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386883, 'name': CreateVM_Task, 'duration_secs': 0.241092} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.393312] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 599.394216] env[62183]: DEBUG oslo_vmware.service [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3eb1c05-3741-4a02-a783-01030d0e0bab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.402536] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.402706] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.403085] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 599.403338] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-254b691a-22cd-452d-8012-d70cb1ee3572 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.409795] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 599.409795] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]523e4e30-8310-3779-643b-64a6499f656b" [ 599.409795] env[62183]: _type = "Task" [ 599.409795] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.416241] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e40b4c-c0c0-44fb-8312-17f64a4bcac3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.425161] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.425541] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 599.425659] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.425803] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.425980] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 599.426251] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20ca8a1d-dbfe-4637-b1c2-a5c689e1efda {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.428718] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7feae5-34a9-4282-b195-2747fc2da3e0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.464131] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301952c3-2aec-45a7-9d08-00cb2fab58ab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.468025] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 599.468025] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 599.468270] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278fd9b2-939a-49be-8d5b-c4427772d4d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.476336] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c888ab9b-9c15-40ef-9ecd-62f1f0753d6a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.480580] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35acf80f-d940-426a-bb95-9bc81c7462aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.494970] env[62183]: DEBUG nova.compute.provider_tree [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.498129] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 599.498129] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]527c644b-c0dc-69fc-0f3d-8ee90be5cc42" [ 599.498129] env[62183]: _type = "Task" [ 599.498129] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.507337] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Preparing fetch location {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 599.507576] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Creating directory with path [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 599.507792] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b505c516-4a63-423b-8f83-a8617664f9d9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.528142] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Created directory with path [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 599.528362] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Fetch image to [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 599.528612] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Downloading image file data b250b36d-7961-4edf-b9ae-4be16b3e5ad2 to [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk on the data store datastore2 {{(pid=62183) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 599.529411] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4456138-4f5f-4e7b-8f4c-3eafddf0ebc2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.536774] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a8bd49-5a7e-451b-9449-4e71c904dfa5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.546255] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df050bef-5d52-4bf5-8edd-50013a436314 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.581020] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38f9285-e5b8-4fa1-bc0b-847037c8462e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.582166] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Releasing lock "refresh_cache-3800f34e-b3b6-4f8f-8df9-6f7266e7f901" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.582406] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 599.582627] env[62183]: DEBUG nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.582752] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.587867] env[62183]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-1e0892b4-47dc-4394-9bbe-dd36bef2dec8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.599115] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.611086] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Downloading image file data b250b36d-7961-4edf-b9ae-4be16b3e5ad2 to the data store datastore2 {{(pid=62183) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 599.669157] env[62183]: DEBUG oslo_vmware.rw_handles [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62183) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 599.733917] env[62183]: INFO nova.scheduler.client.report [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted allocations for instance 01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff [ 599.789330] env[62183]: INFO nova.compute.manager [-] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Took 1.02 seconds to deallocate network for instance. [ 599.792308] env[62183]: DEBUG nova.compute.claims [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.792560] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.998950] env[62183]: DEBUG nova.scheduler.client.report [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 600.102107] env[62183]: DEBUG nova.network.neutron [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.163851] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.163851] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.243715] env[62183]: DEBUG oslo_concurrency.lockutils [None req-14bf5cfe-7845-49aa-bde5-53818ce3234a tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "01ab1e03-4b2e-4ee4-82f9-1fa5c90aedff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.435s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.411431] env[62183]: DEBUG oslo_vmware.rw_handles [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Completed reading data from the image iterator. {{(pid=62183) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 600.411632] env[62183]: DEBUG oslo_vmware.rw_handles [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 600.466848] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Downloaded image file data b250b36d-7961-4edf-b9ae-4be16b3e5ad2 to vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk on the data store datastore2 {{(pid=62183) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 600.468997] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Caching image {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 600.469277] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Copying Virtual Disk [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk to [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 600.469730] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d77ff399-5dcf-490e-8cf1-3ae32393712e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.478688] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 600.478688] env[62183]: value = "task-1386884" [ 600.478688] env[62183]: _type = "Task" [ 600.478688] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.486706] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.502744] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.503458] env[62183]: ERROR nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a6802bb8-ad1d-4496-b8c6-277231254e26, please check neutron logs for more information. [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Traceback (most recent call last): [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self.driver.spawn(context, instance, image_meta, [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] vm_ref = self.build_virtual_machine(instance, [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.503458] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] for vif in network_info: [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] return self._sync_wrapper(fn, *args, **kwargs) [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self.wait() [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self[:] = self._gt.wait() [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] return self._exit_event.wait() [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] result = hub.switch() [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.503774] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] return self.greenlet.switch() [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] result = function(*args, **kwargs) [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] return func(*args, **kwargs) [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] raise e [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] nwinfo = self.network_api.allocate_for_instance( [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] created_port_ids = self._update_ports_for_instance( [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] with excutils.save_and_reraise_exception(): [ 600.504113] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] self.force_reraise() [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] raise self.value [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] updated_port = self._update_port( [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] _ensure_no_port_binding_failure(port) [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] raise exception.PortBindingFailed(port_id=port['id']) [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] nova.exception.PortBindingFailed: Binding failed for port a6802bb8-ad1d-4496-b8c6-277231254e26, please check neutron logs for more information. [ 600.504462] env[62183]: ERROR nova.compute.manager [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] [ 600.504748] env[62183]: DEBUG nova.compute.utils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Binding failed for port a6802bb8-ad1d-4496-b8c6-277231254e26, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 600.505295] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.207s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.508282] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Build of instance 2aa13e44-de7e-476a-aa3d-0675d38b5ef2 was re-scheduled: Binding failed for port a6802bb8-ad1d-4496-b8c6-277231254e26, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 600.508699] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 600.508924] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Acquiring lock "refresh_cache-2aa13e44-de7e-476a-aa3d-0675d38b5ef2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.509080] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Acquired lock "refresh_cache-2aa13e44-de7e-476a-aa3d-0675d38b5ef2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.509242] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.605068] env[62183]: INFO nova.compute.manager [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3800f34e-b3b6-4f8f-8df9-6f7266e7f901] Took 1.02 seconds to deallocate network for instance. [ 600.669043] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.669043] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 600.669199] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Rebuilding the list of instances to heal {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 600.747636] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 600.990635] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386884, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.039069] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.173131] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 601.173131] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 601.179051] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.211549] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "refresh_cache-a11cfd43-7516-4f98-8484-9ba2a0c5276d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.211690] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquired lock "refresh_cache-a11cfd43-7516-4f98-8484-9ba2a0c5276d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.211832] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Forcefully refreshing network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 601.212017] env[62183]: DEBUG nova.objects.instance [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lazy-loading 'info_cache' on Instance uuid a11cfd43-7516-4f98-8484-9ba2a0c5276d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 601.267276] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.428093] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08cc575-9aba-4002-9b3d-f5b8e6e00acb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.437473] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f6ff35-feb0-48bc-98c0-01d4d87c4084 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.476476] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca06c9b-8193-481a-abdd-ae0e80c30a8e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.485837] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9dff3b-b0c3-4cc5-aad9-763ae4457f29 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.492839] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386884, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67985} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.493466] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Copied Virtual Disk [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk to [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 601.493694] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Deleting the datastore file [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/tmp-sparse.vmdk {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 601.493927] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da537048-ddbf-450f-99a7-cc17c46cd751 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.503403] env[62183]: DEBUG nova.compute.provider_tree [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.510378] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 601.510378] env[62183]: value = "task-1386885" [ 601.510378] env[62183]: _type = "Task" [ 601.510378] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.519733] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386885, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.633971] env[62183]: INFO nova.scheduler.client.report [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Deleted allocations for instance 3800f34e-b3b6-4f8f-8df9-6f7266e7f901 [ 601.684503] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Releasing lock "refresh_cache-2aa13e44-de7e-476a-aa3d-0675d38b5ef2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.684760] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 601.684966] env[62183]: DEBUG nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.685148] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.740150] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.990913] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "82816947-cb4f-4ad7-ad44-1403db312cf8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.990913] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "82816947-cb4f-4ad7-ad44-1403db312cf8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.006747] env[62183]: DEBUG nova.scheduler.client.report [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 602.020799] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386885, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023076} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.021112] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 602.021358] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Moving file from [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb/b250b36d-7961-4edf-b9ae-4be16b3e5ad2 to [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2. {{(pid=62183) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 602.021608] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-582ea642-1586-4ece-bb02-b904f41cbc44 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.028913] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 602.028913] env[62183]: value = "task-1386886" [ 602.028913] env[62183]: _type = "Task" [ 602.028913] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.036587] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386886, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.144596] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0125ff3c-d1d7-4fa3-aff5-bb47f4a036bc tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "3800f34e-b3b6-4f8f-8df9-6f7266e7f901" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.434s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.239320] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.244463] env[62183]: DEBUG nova.network.neutron [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.489556] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Acquiring lock "a1cabc70-e6c0-4c9b-8302-6ed8805f2182" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.489939] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Lock "a1cabc70-e6c0-4c9b-8302-6ed8805f2182" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.516661] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.011s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.517312] env[62183]: ERROR nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bc18163b-b86f-4c54-9b0e-1f68a35f444e, please check neutron logs for more information. [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Traceback (most recent call last): [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self.driver.spawn(context, instance, image_meta, [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] vm_ref = self.build_virtual_machine(instance, [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.517312] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] for vif in network_info: [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] return self._sync_wrapper(fn, *args, **kwargs) [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self.wait() [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self[:] = self._gt.wait() [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] return self._exit_event.wait() [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] current.throw(*self._exc) [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.517679] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] result = function(*args, **kwargs) [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] return func(*args, **kwargs) [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] raise e [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] nwinfo = self.network_api.allocate_for_instance( [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] created_port_ids = self._update_ports_for_instance( [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] with excutils.save_and_reraise_exception(): [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] self.force_reraise() [ 602.518041] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.518963] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] raise self.value [ 602.518963] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.518963] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] updated_port = self._update_port( [ 602.518963] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.518963] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] _ensure_no_port_binding_failure(port) [ 602.518963] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.518963] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] raise exception.PortBindingFailed(port_id=port['id']) [ 602.518963] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] nova.exception.PortBindingFailed: Binding failed for port bc18163b-b86f-4c54-9b0e-1f68a35f444e, please check neutron logs for more information. [ 602.518963] env[62183]: ERROR nova.compute.manager [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] [ 602.518963] env[62183]: DEBUG nova.compute.utils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Binding failed for port bc18163b-b86f-4c54-9b0e-1f68a35f444e, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.519739] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Build of instance 4c9534c7-5498-46a1-b6c2-c93857031078 was re-scheduled: Binding failed for port bc18163b-b86f-4c54-9b0e-1f68a35f444e, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 602.520295] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 602.520554] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Acquiring lock "refresh_cache-4c9534c7-5498-46a1-b6c2-c93857031078" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.520721] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Acquired lock "refresh_cache-4c9534c7-5498-46a1-b6c2-c93857031078" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.520894] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.522081] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.608s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.539620] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386886, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024653} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.539867] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] File moved {{(pid=62183) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 602.540435] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Cleaning up location [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 602.540645] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Deleting the datastore file [datastore2] vmware_temp/f1632b17-6eae-4d8b-ace0-2542fc31b2fb {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 602.540914] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19127ba5-2910-4744-b375-af7500b72eff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.550012] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 602.550012] env[62183]: value = "task-1386887" [ 602.550012] env[62183]: _type = "Task" [ 602.550012] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.559814] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.647241] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.750941] env[62183]: INFO nova.compute.manager [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] [instance: 2aa13e44-de7e-476a-aa3d-0675d38b5ef2] Took 1.07 seconds to deallocate network for instance. [ 602.821771] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.047025] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.060380] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022804} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.060380] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 603.062905] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f7e0fbc-e607-4fc7-b85a-0490a4c8ddff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.068594] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 603.068594] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]529692da-0502-5ec5-d2f3-d9a422d37d83" [ 603.068594] env[62183]: _type = "Task" [ 603.068594] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.076364] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529692da-0502-5ec5-d2f3-d9a422d37d83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.141622] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.169511] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.324103] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Releasing lock "refresh_cache-a11cfd43-7516-4f98-8484-9ba2a0c5276d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.324335] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Updated the network info_cache for instance {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 603.326846] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.327239] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.327708] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.327871] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.328060] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.328233] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.328363] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 603.328807] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.428571] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a54ffe9-7406-4901-b479-5e336332fa9f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.436603] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4dce2fe-614f-4deb-9157-fca46534975a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.467309] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba45e65d-8dcc-4ec7-9121-da7bffa32d21 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.475346] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144bd299-0946-4890-970e-8d53940b11fe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.489816] env[62183]: DEBUG nova.compute.provider_tree [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.578619] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529692da-0502-5ec5-d2f3-d9a422d37d83, 'name': SearchDatastore_Task, 'duration_secs': 0.008169} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.578963] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.579183] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 603.579433] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cac5fdce-e16c-40f9-ad03-86e06e490f8a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.586124] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 603.586124] env[62183]: value = "task-1386888" [ 603.586124] env[62183]: _type = "Task" [ 603.586124] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.593900] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386888, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.644871] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Releasing lock "refresh_cache-4c9534c7-5498-46a1-b6c2-c93857031078" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.645021] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 603.645310] env[62183]: DEBUG nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.645415] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.664987] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.784432] env[62183]: INFO nova.scheduler.client.report [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Deleted allocations for instance 2aa13e44-de7e-476a-aa3d-0675d38b5ef2 [ 603.833892] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.994188] env[62183]: DEBUG nova.scheduler.client.report [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 604.096316] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386888, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483839} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.096316] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 604.096316] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 604.096746] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38456800-a3b9-41bf-ae51-cdfb1fc0288b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.102695] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 604.102695] env[62183]: value = "task-1386889" [ 604.102695] env[62183]: _type = "Task" [ 604.102695] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.110646] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386889, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.169440] env[62183]: DEBUG nova.network.neutron [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.295821] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d3851b1-f871-4dca-930d-b260e44f7c94 tempest-VolumesAssistedSnapshotsTest-152773177 tempest-VolumesAssistedSnapshotsTest-152773177-project-member] Lock "2aa13e44-de7e-476a-aa3d-0675d38b5ef2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.889s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.410020] env[62183]: DEBUG nova.compute.manager [None req-48192106-502b-49c4-9e27-360e6ca13f17 tempest-ServerDiagnosticsV248Test-2011752479 tempest-ServerDiagnosticsV248Test-2011752479-project-admin] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 604.410876] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de63307-ee7d-4702-a47b-d48d55b12854 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.418033] env[62183]: INFO nova.compute.manager [None req-48192106-502b-49c4-9e27-360e6ca13f17 tempest-ServerDiagnosticsV248Test-2011752479 tempest-ServerDiagnosticsV248Test-2011752479-project-admin] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Retrieving diagnostics [ 604.419020] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e06a389-5cee-4bd0-86c8-f46e1c94e6c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.501916] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.502589] env[62183]: ERROR nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 62394ec9-e861-48dd-bca0-75b71c5f97f0, please check neutron logs for more information. [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Traceback (most recent call last): [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self.driver.spawn(context, instance, image_meta, [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] vm_ref = self.build_virtual_machine(instance, [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.502589] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] for vif in network_info: [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] return self._sync_wrapper(fn, *args, **kwargs) [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self.wait() [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self[:] = self._gt.wait() [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] return self._exit_event.wait() [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] current.throw(*self._exc) [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.502858] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] result = function(*args, **kwargs) [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] return func(*args, **kwargs) [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] raise e [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] nwinfo = self.network_api.allocate_for_instance( [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] created_port_ids = self._update_ports_for_instance( [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] with excutils.save_and_reraise_exception(): [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] self.force_reraise() [ 604.503160] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.503491] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] raise self.value [ 604.503491] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.503491] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] updated_port = self._update_port( [ 604.503491] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.503491] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] _ensure_no_port_binding_failure(port) [ 604.503491] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.503491] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] raise exception.PortBindingFailed(port_id=port['id']) [ 604.503491] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] nova.exception.PortBindingFailed: Binding failed for port 62394ec9-e861-48dd-bca0-75b71c5f97f0, please check neutron logs for more information. [ 604.503491] env[62183]: ERROR nova.compute.manager [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] [ 604.503491] env[62183]: DEBUG nova.compute.utils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Binding failed for port 62394ec9-e861-48dd-bca0-75b71c5f97f0, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.504499] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.120s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.506058] env[62183]: INFO nova.compute.claims [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.508801] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Build of instance 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5 was re-scheduled: Binding failed for port 62394ec9-e861-48dd-bca0-75b71c5f97f0, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.509242] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.509483] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.509630] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.509788] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.613753] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386889, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080567} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.613978] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 604.614760] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887d0ecf-cf9b-400b-84e9-f60b0e5b213b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.634069] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 604.634332] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5199b8c-914b-45b7-9f2f-c7218625f3b0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.652915] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 604.652915] env[62183]: value = "task-1386890" [ 604.652915] env[62183]: _type = "Task" [ 604.652915] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.660595] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386890, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.672260] env[62183]: INFO nova.compute.manager [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] [instance: 4c9534c7-5498-46a1-b6c2-c93857031078] Took 1.03 seconds to deallocate network for instance. [ 604.801283] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 604.831553] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "3cb4be60-b626-4cc9-91eb-2bd9b778f153" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.831795] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "3cb4be60-b626-4cc9-91eb-2bd9b778f153" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.038734] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.147072] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.165894] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386890, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.324853] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.617427] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquiring lock "a11cfd43-7516-4f98-8484-9ba2a0c5276d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.617670] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "a11cfd43-7516-4f98-8484-9ba2a0c5276d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.617867] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquiring lock "a11cfd43-7516-4f98-8484-9ba2a0c5276d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.618069] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "a11cfd43-7516-4f98-8484-9ba2a0c5276d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.618223] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "a11cfd43-7516-4f98-8484-9ba2a0c5276d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.623646] env[62183]: INFO nova.compute.manager [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Terminating instance [ 605.625232] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquiring lock "refresh_cache-a11cfd43-7516-4f98-8484-9ba2a0c5276d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.625394] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquired lock "refresh_cache-a11cfd43-7516-4f98-8484-9ba2a0c5276d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.625555] env[62183]: DEBUG nova.network.neutron [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.650258] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.650494] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.650673] env[62183]: DEBUG nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.650834] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.662876] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386890, 'name': ReconfigVM_Task, 'duration_secs': 0.553568} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.664985] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Reconfigured VM instance instance-00000011 to attach disk [datastore2] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 605.665755] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76f1c47d-8586-4872-9dac-d24e5db9e959 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.672594] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 605.672594] env[62183]: value = "task-1386891" [ 605.672594] env[62183]: _type = "Task" [ 605.672594] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.673304] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.686940] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386891, 'name': Rename_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.699186] env[62183]: INFO nova.scheduler.client.report [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Deleted allocations for instance 4c9534c7-5498-46a1-b6c2-c93857031078 [ 605.874172] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e00a99-8d80-4bdf-b2b6-657cf810a681 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.881850] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627e8e88-aed5-428e-8f56-2106e9b1a510 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.911047] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b91edfa-7ce8-4d19-a37e-a35427a314b1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.918156] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87fc618-070e-4ff8-a518-b599e854a36a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.931012] env[62183]: DEBUG nova.compute.provider_tree [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.143073] env[62183]: DEBUG nova.network.neutron [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.180343] env[62183]: DEBUG nova.network.neutron [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.184812] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386891, 'name': Rename_Task, 'duration_secs': 0.125508} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.186263] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 606.186263] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06e41f35-752d-460d-95eb-2170f7a57711 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.194371] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 606.194371] env[62183]: value = "task-1386892" [ 606.194371] env[62183]: _type = "Task" [ 606.194371] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.200241] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.212019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b5e1ebe9-8732-4c65-8a1a-846c73ff4158 tempest-AttachInterfacesUnderV243Test-1975942314 tempest-AttachInterfacesUnderV243Test-1975942314-project-member] Lock "4c9534c7-5498-46a1-b6c2-c93857031078" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.603s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.237793] env[62183]: DEBUG nova.network.neutron [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.434181] env[62183]: DEBUG nova.scheduler.client.report [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 606.686418] env[62183]: INFO nova.compute.manager [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5] Took 1.04 seconds to deallocate network for instance. [ 606.706906] env[62183]: DEBUG oslo_vmware.api [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386892, 'name': PowerOnVM_Task, 'duration_secs': 0.409133} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.710267] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 606.710507] env[62183]: DEBUG nova.compute.manager [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 606.710887] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.713726] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fade2883-2355-461d-b528-78ce4681d4b0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.742280] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Releasing lock "refresh_cache-a11cfd43-7516-4f98-8484-9ba2a0c5276d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.742280] env[62183]: DEBUG nova.compute.manager [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.742280] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 606.742280] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcda3920-fa7c-4e55-a416-1bd53adcbde1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.748944] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 606.749150] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-536ad8b4-cc33-41f3-8758-5654bbbc4179 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.754877] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 606.754877] env[62183]: value = "task-1386893" [ 606.754877] env[62183]: _type = "Task" [ 606.754877] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.764056] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.941476] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.941891] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.945129] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.694s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.946458] env[62183]: INFO nova.compute.claims [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.239568] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.252368] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.265442] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386893, 'name': PowerOffVM_Task, 'duration_secs': 0.123964} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.265693] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 607.265858] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 607.266133] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf90b462-43db-42a1-8857-8a715c3b95d4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.290714] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 607.290714] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 607.290901] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Deleting the datastore file [datastore1] a11cfd43-7516-4f98-8484-9ba2a0c5276d {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 607.291674] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81d5a5cc-7074-44c9-94ea-877bf93d9068 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.298384] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for the task: (returnval){ [ 607.298384] env[62183]: value = "task-1386895" [ 607.298384] env[62183]: _type = "Task" [ 607.298384] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 607.310787] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.450629] env[62183]: DEBUG nova.compute.utils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.460344] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.460344] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 607.547531] env[62183]: DEBUG nova.policy [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '145f4f1d5ba14efe87430102c176777a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '276875a2349a42d3ae4fec83244ba8f9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.735214] env[62183]: INFO nova.scheduler.client.report [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Deleted allocations for instance 85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5 [ 607.744980] env[62183]: INFO nova.compute.manager [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Rebuilding instance [ 607.806342] env[62183]: DEBUG nova.compute.manager [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 607.807796] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306bc521-dd54-406b-8152-e4a188eed54d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.813548] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.960514] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.250170] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ee2b2a5a-4b74-4d94-996b-8508e6ea1d00 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "85de5dc1-8e94-4f65-ab41-a80e5ff9c8e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.629s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.307967] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Successfully created port: 92cf027c-9f9f-4b98-b960-273b2ab787e1 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.315986] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.323129] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 608.323594] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fba9b0da-395e-4db1-b161-0b9b6b7601d4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.330387] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 608.330387] env[62183]: value = "task-1386896" [ 608.330387] env[62183]: _type = "Task" [ 608.330387] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.338648] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386896, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.344399] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dab4773-6ed8-4b4c-bd77-a1f5913bfc9e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.351321] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf1f1a0-512a-4815-8143-f7f03895bf00 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.392018] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcad331a-ec2f-4fe7-b665-ccad89f6e588 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.401362] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c13b8d7-c99a-4edb-a183-330173fc10cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.418575] env[62183]: DEBUG nova.compute.provider_tree [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.466616] env[62183]: INFO nova.virt.block_device [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Booting with volume 7faeed94-b2a8-463a-8f4b-56e80a8ab919 at /dev/sda [ 608.525376] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a190ac0a-5a85-4ede-9854-591552a6f897 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.533933] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ed7794-f08f-483c-b14c-873aa6647c96 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.558783] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3966def7-3f6d-4621-99b9-48755ea9bba7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.566351] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c73aba2-9ff2-4efb-a733-870897039523 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.588939] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d8d235-21d5-4f31-a712-7b64bab8e47a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.595204] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1e214c-83a0-477e-9fd1-28dc0b42bf5b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.609452] env[62183]: DEBUG nova.virt.block_device [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Updating existing volume attachment record: d7fee083-7084-422e-8f00-4948e94c3946 {{(pid=62183) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 608.756161] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.810598] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.840801] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386896, 'name': PowerOffVM_Task, 'duration_secs': 0.181601} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 608.841109] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 608.841433] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.842318] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d99835-c077-481f-87ac-2d919f0ce67d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.848825] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 608.849075] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-984155af-9fca-49d9-badc-557946368f0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.876437] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 608.876606] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 608.876792] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Deleting the datastore file [datastore2] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 608.877134] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0858cb4c-da2f-476b-8e5e-d90465df6157 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.884585] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 608.884585] env[62183]: value = "task-1386898" [ 608.884585] env[62183]: _type = "Task" [ 608.884585] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 608.894613] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 608.922094] env[62183]: DEBUG nova.scheduler.client.report [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 609.278686] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.312628] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.396865] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101404} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.396865] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 609.396865] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 609.396865] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 609.428685] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.430588] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 609.433159] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.029s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.435093] env[62183]: INFO nova.compute.claims [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.803139] env[62183]: DEBUG nova.compute.manager [req-bb00ee0c-38e3-47c1-a46f-ac5c75054961 req-85e6b279-38df-47a9-9ae2-9753b046c3ce service nova] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Received event network-changed-92cf027c-9f9f-4b98-b960-273b2ab787e1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.803139] env[62183]: DEBUG nova.compute.manager [req-bb00ee0c-38e3-47c1-a46f-ac5c75054961 req-85e6b279-38df-47a9-9ae2-9753b046c3ce service nova] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Refreshing instance network info cache due to event network-changed-92cf027c-9f9f-4b98-b960-273b2ab787e1. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 609.803139] env[62183]: DEBUG oslo_concurrency.lockutils [req-bb00ee0c-38e3-47c1-a46f-ac5c75054961 req-85e6b279-38df-47a9-9ae2-9753b046c3ce service nova] Acquiring lock "refresh_cache-3ec662c7-940e-43fa-bbea-f3bc3243adf4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.803139] env[62183]: DEBUG oslo_concurrency.lockutils [req-bb00ee0c-38e3-47c1-a46f-ac5c75054961 req-85e6b279-38df-47a9-9ae2-9753b046c3ce service nova] Acquired lock "refresh_cache-3ec662c7-940e-43fa-bbea-f3bc3243adf4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.803139] env[62183]: DEBUG nova.network.neutron [req-bb00ee0c-38e3-47c1-a46f-ac5c75054961 req-85e6b279-38df-47a9-9ae2-9753b046c3ce service nova] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Refreshing network info cache for port 92cf027c-9f9f-4b98-b960-273b2ab787e1 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 609.816842] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.942022] env[62183]: DEBUG nova.compute.utils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.944274] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.944613] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 610.093039] env[62183]: DEBUG nova.policy [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a727a66bfbf43639d3a4c388c5f1753', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e21ad21261534986a51e86e8b0b5a7a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 610.209561] env[62183]: ERROR nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 92cf027c-9f9f-4b98-b960-273b2ab787e1, please check neutron logs for more information. [ 610.209561] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.209561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.209561] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.209561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.209561] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.209561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.209561] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.209561] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.209561] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 610.209561] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.209561] env[62183]: ERROR nova.compute.manager raise self.value [ 610.209561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.209561] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.209561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.209561] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.210872] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.210872] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.210872] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 92cf027c-9f9f-4b98-b960-273b2ab787e1, please check neutron logs for more information. [ 610.210872] env[62183]: ERROR nova.compute.manager [ 610.210872] env[62183]: Traceback (most recent call last): [ 610.210872] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.210872] env[62183]: listener.cb(fileno) [ 610.210872] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.210872] env[62183]: result = function(*args, **kwargs) [ 610.210872] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.210872] env[62183]: return func(*args, **kwargs) [ 610.210872] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.210872] env[62183]: raise e [ 610.210872] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.210872] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 610.210872] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.210872] env[62183]: created_port_ids = self._update_ports_for_instance( [ 610.210872] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.210872] env[62183]: with excutils.save_and_reraise_exception(): [ 610.210872] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.210872] env[62183]: self.force_reraise() [ 610.210872] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.210872] env[62183]: raise self.value [ 610.210872] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.210872] env[62183]: updated_port = self._update_port( [ 610.210872] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.210872] env[62183]: _ensure_no_port_binding_failure(port) [ 610.210872] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.210872] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.211902] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 92cf027c-9f9f-4b98-b960-273b2ab787e1, please check neutron logs for more information. [ 610.211902] env[62183]: Removing descriptor: 16 [ 610.935689] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Successfully created port: 48058051-c983-4135-8170-3395eeadc4e3 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 610.942214] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 610.947873] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 610.948441] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.948653] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.948803] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.948978] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.950061] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.950061] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.950061] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.950061] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.950061] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.950362] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.950362] env[62183]: DEBUG nova.virt.hardware [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.954503] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c649d5-4e54-4c6c-b3a0-2a5e95dbbe3f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.966131] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e4ef2c-7c9a-4bb8-94e7-c3f4931d0672 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.971268] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.982881] env[62183]: ERROR nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 92cf027c-9f9f-4b98-b960-273b2ab787e1, please check neutron logs for more information. [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Traceback (most recent call last): [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] yield resources [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self.driver.spawn(context, instance, image_meta, [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] vm_ref = self.build_virtual_machine(instance, [ 610.982881] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] for vif in network_info: [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] return self._sync_wrapper(fn, *args, **kwargs) [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self.wait() [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self[:] = self._gt.wait() [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] return self._exit_event.wait() [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 610.983767] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] current.throw(*self._exc) [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] result = function(*args, **kwargs) [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] return func(*args, **kwargs) [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] raise e [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] nwinfo = self.network_api.allocate_for_instance( [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] created_port_ids = self._update_ports_for_instance( [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] with excutils.save_and_reraise_exception(): [ 610.984255] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self.force_reraise() [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] raise self.value [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] updated_port = self._update_port( [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] _ensure_no_port_binding_failure(port) [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] raise exception.PortBindingFailed(port_id=port['id']) [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] nova.exception.PortBindingFailed: Binding failed for port 92cf027c-9f9f-4b98-b960-273b2ab787e1, please check neutron logs for more information. [ 610.984777] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] [ 610.984777] env[62183]: INFO nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Terminating instance [ 610.986732] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.986974] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.987151] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.987333] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.987473] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.987616] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.987817] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.987970] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.988151] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.988310] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.988478] env[62183]: DEBUG nova.virt.hardware [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.992193] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef40e65f-e68f-4bef-ac75-6b6722833081 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.995838] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Acquiring lock "refresh_cache-3ec662c7-940e-43fa-bbea-f3bc3243adf4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.000618] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9a684b-6ffb-4d6e-8675-eba71f505a7f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.015331] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 611.022044] env[62183]: DEBUG oslo.service.loopingcall [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.025015] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 611.026287] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd36897f-6054-471b-bdf9-001273d0f05b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.042074] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 611.042074] env[62183]: value = "task-1386899" [ 611.042074] env[62183]: _type = "Task" [ 611.042074] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.049520] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386899, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.171796] env[62183]: DEBUG nova.network.neutron [req-bb00ee0c-38e3-47c1-a46f-ac5c75054961 req-85e6b279-38df-47a9-9ae2-9753b046c3ce service nova] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.404094] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec63d65-f274-403a-9e36-0604a6fcaf37 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.410858] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b219e0-c997-4e92-95fc-b5c9584b458e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.450989] env[62183]: DEBUG nova.network.neutron [req-bb00ee0c-38e3-47c1-a46f-ac5c75054961 req-85e6b279-38df-47a9-9ae2-9753b046c3ce service nova] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.458061] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e26f4e-b87b-4d00-a522-850d185f1787 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.468786] env[62183]: DEBUG oslo_vmware.api [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Task: {'id': task-1386895, 'name': DeleteDatastoreFile_Task, 'duration_secs': 3.727524} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.471026] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 611.471688] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 611.471688] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 611.471688] env[62183]: INFO nova.compute.manager [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Took 4.73 seconds to destroy the instance on the hypervisor. [ 611.471835] env[62183]: DEBUG oslo.service.loopingcall [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.472275] env[62183]: DEBUG nova.compute.manager [-] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.472378] env[62183]: DEBUG nova.network.neutron [-] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.475026] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecb6616-6b97-4a4d-8bcc-42427edfd2b7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.490813] env[62183]: DEBUG nova.compute.provider_tree [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.514910] env[62183]: DEBUG nova.network.neutron [-] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.551213] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386899, 'name': CreateVM_Task} progress is 15%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.965037] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.967335] env[62183]: DEBUG oslo_concurrency.lockutils [req-bb00ee0c-38e3-47c1-a46f-ac5c75054961 req-85e6b279-38df-47a9-9ae2-9753b046c3ce service nova] Releasing lock "refresh_cache-3ec662c7-940e-43fa-bbea-f3bc3243adf4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.967899] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Acquired lock "refresh_cache-3ec662c7-940e-43fa-bbea-f3bc3243adf4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.968089] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.990339] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.990612] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.990767] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.990945] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.991217] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.991490] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.991733] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.991894] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.992230] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.992230] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.992451] env[62183]: DEBUG nova.virt.hardware [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.993581] env[62183]: DEBUG nova.scheduler.client.report [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 611.997325] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df046bc7-c9e0-468f-a682-3905f8dfb5e7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.010044] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcca825b-6889-4a72-9dac-0c553cb1e183 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.026455] env[62183]: DEBUG nova.network.neutron [-] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.051671] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386899, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.084881] env[62183]: DEBUG nova.compute.manager [req-832ded9e-dc71-412d-92a6-cbf8e2ca840f req-053c54f7-60d8-431f-b2b4-eaae53df1d76 service nova] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Received event network-vif-deleted-92cf027c-9f9f-4b98-b960-273b2ab787e1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.360225] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "f25ef564-b649-4557-9847-b5d994079d20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.360480] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "f25ef564-b649-4557-9847-b5d994079d20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.500170] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.502533] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.069s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.503833] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 612.506842] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.726s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.509538] env[62183]: INFO nova.compute.claims [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.529552] env[62183]: INFO nova.compute.manager [-] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Took 1.06 seconds to deallocate network for instance. [ 612.555116] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386899, 'name': CreateVM_Task, 'duration_secs': 1.205629} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.555448] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 612.555893] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.556108] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.556455] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 612.556762] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5593dfbe-ea04-486e-bc0d-579a72247bdb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.561451] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 612.561451] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52655397-1232-17a8-a4f2-753457eaf6fb" [ 612.561451] env[62183]: _type = "Task" [ 612.561451] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.576063] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52655397-1232-17a8-a4f2-753457eaf6fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.653126] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.902963] env[62183]: ERROR nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 48058051-c983-4135-8170-3395eeadc4e3, please check neutron logs for more information. [ 612.902963] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 612.902963] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.902963] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 612.902963] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.902963] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 612.902963] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.902963] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 612.902963] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.902963] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 612.902963] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.902963] env[62183]: ERROR nova.compute.manager raise self.value [ 612.902963] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.902963] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 612.902963] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.902963] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 612.903412] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.903412] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 612.903412] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 48058051-c983-4135-8170-3395eeadc4e3, please check neutron logs for more information. [ 612.903412] env[62183]: ERROR nova.compute.manager [ 612.903412] env[62183]: Traceback (most recent call last): [ 612.903412] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 612.903412] env[62183]: listener.cb(fileno) [ 612.903412] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.903412] env[62183]: result = function(*args, **kwargs) [ 612.903412] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 612.903412] env[62183]: return func(*args, **kwargs) [ 612.903412] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.903412] env[62183]: raise e [ 612.903412] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.903412] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 612.903412] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.903412] env[62183]: created_port_ids = self._update_ports_for_instance( [ 612.903412] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.903412] env[62183]: with excutils.save_and_reraise_exception(): [ 612.903412] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.903412] env[62183]: self.force_reraise() [ 612.903412] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.903412] env[62183]: raise self.value [ 612.903412] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.903412] env[62183]: updated_port = self._update_port( [ 612.903412] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.903412] env[62183]: _ensure_no_port_binding_failure(port) [ 612.903412] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.903412] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 612.904035] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 48058051-c983-4135-8170-3395eeadc4e3, please check neutron logs for more information. [ 612.904035] env[62183]: Removing descriptor: 14 [ 612.904035] env[62183]: ERROR nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 48058051-c983-4135-8170-3395eeadc4e3, please check neutron logs for more information. [ 612.904035] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Traceback (most recent call last): [ 612.904035] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 612.904035] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] yield resources [ 612.904035] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.904035] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self.driver.spawn(context, instance, image_meta, [ 612.904035] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 612.904035] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.904035] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.904035] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] vm_ref = self.build_virtual_machine(instance, [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] for vif in network_info: [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] return self._sync_wrapper(fn, *args, **kwargs) [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self.wait() [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self[:] = self._gt.wait() [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] return self._exit_event.wait() [ 612.904347] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] result = hub.switch() [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] return self.greenlet.switch() [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] result = function(*args, **kwargs) [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] return func(*args, **kwargs) [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] raise e [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] nwinfo = self.network_api.allocate_for_instance( [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.904668] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] created_port_ids = self._update_ports_for_instance( [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] with excutils.save_and_reraise_exception(): [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self.force_reraise() [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] raise self.value [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] updated_port = self._update_port( [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] _ensure_no_port_binding_failure(port) [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.905018] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] raise exception.PortBindingFailed(port_id=port['id']) [ 612.905319] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] nova.exception.PortBindingFailed: Binding failed for port 48058051-c983-4135-8170-3395eeadc4e3, please check neutron logs for more information. [ 612.905319] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] [ 612.905319] env[62183]: INFO nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Terminating instance [ 612.909915] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Acquiring lock "refresh_cache-79c10e4e-98ab-4259-8ff5-72bfa39860be" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.909915] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Acquired lock "refresh_cache-79c10e4e-98ab-4259-8ff5-72bfa39860be" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.909915] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.015756] env[62183]: DEBUG nova.compute.utils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 613.024102] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 613.027053] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 613.042581] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.074580] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52655397-1232-17a8-a4f2-753457eaf6fb, 'name': SearchDatastore_Task, 'duration_secs': 0.015605} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.074950] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.075356] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 613.075356] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.075469] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.075587] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 613.075862] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fccd2940-99e0-42b1-8c4c-6a85b3d7a2ae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.090256] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 613.090589] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 613.091342] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09added0-02e4-4dff-8410-8b7ca1ba1125 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.098990] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 613.098990] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52b8e198-5131-ff86-0aad-ba193ee97c9a" [ 613.098990] env[62183]: _type = "Task" [ 613.098990] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.109341] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b8e198-5131-ff86-0aad-ba193ee97c9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.110437] env[62183]: DEBUG nova.policy [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf20d12101324dc79087160458adfdad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e8a33c5c8f0d45198e736bcad07c5206', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 613.156908] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Releasing lock "refresh_cache-3ec662c7-940e-43fa-bbea-f3bc3243adf4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.157942] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 613.157942] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-661a07d8-5593-4808-8acd-65fec4179b08 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.172022] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6437e17-dd79-4b95-a921-a2aa01949f9e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.196287] env[62183]: WARNING nova.virt.vmwareapi.driver [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 3ec662c7-940e-43fa-bbea-f3bc3243adf4 could not be found. [ 613.196593] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 613.196957] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b60a8657-9837-4a55-895b-25e360c0eaa9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.204873] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbe9ebb-d6e7-4195-a085-812c9de2f4fd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.230902] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3ec662c7-940e-43fa-bbea-f3bc3243adf4 could not be found. [ 613.231283] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.231397] env[62183]: INFO nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Took 0.07 seconds to destroy the instance on the hypervisor. [ 613.231771] env[62183]: DEBUG oslo.service.loopingcall [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.231979] env[62183]: DEBUG nova.compute.manager [-] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.232330] env[62183]: DEBUG nova.network.neutron [-] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.254131] env[62183]: DEBUG nova.network.neutron [-] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.439427] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.524570] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 613.573349] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.615675] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b8e198-5131-ff86-0aad-ba193ee97c9a, 'name': SearchDatastore_Task, 'duration_secs': 0.009664} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.618608] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50dd684b-3d54-4519-8c2d-c14b73981985 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.624257] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 613.624257] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]528f1add-e9b0-5bd5-c82a-92062b345bc4" [ 613.624257] env[62183]: _type = "Task" [ 613.624257] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.634798] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528f1add-e9b0-5bd5-c82a-92062b345bc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.681943] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Successfully created port: 0b1c71e3-677d-4088-977d-208deb7610fe {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 613.757276] env[62183]: DEBUG nova.network.neutron [-] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.955957] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3eba05-1010-440d-9f03-760a23407444 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.967094] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb39485-7f69-4d51-a01f-66f2fad2de0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.004933] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6db0a17-79d2-440d-9267-449522e68331 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.013901] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa3a163-3a46-4e83-9fa3-359481943fea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.033514] env[62183]: DEBUG nova.compute.provider_tree [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.078847] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Releasing lock "refresh_cache-79c10e4e-98ab-4259-8ff5-72bfa39860be" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.078847] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.078847] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 614.078847] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70b2efd4-b024-4d5e-9aa1-88c81728c6ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.088638] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cb26fd-b353-4566-a06f-71bb6edc9510 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.115819] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 79c10e4e-98ab-4259-8ff5-72bfa39860be could not be found. [ 614.116052] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 614.116319] env[62183]: INFO nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Took 0.04 seconds to destroy the instance on the hypervisor. [ 614.116581] env[62183]: DEBUG oslo.service.loopingcall [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.118241] env[62183]: DEBUG nova.compute.manager [-] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.118366] env[62183]: DEBUG nova.network.neutron [-] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.124038] env[62183]: DEBUG nova.compute.manager [req-1d306f7b-acff-4f6c-bfdd-9902c147a5ab req-0b8cf4ba-587d-49a3-ab46-58a427dfde5d service nova] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Received event network-changed-48058051-c983-4135-8170-3395eeadc4e3 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 614.124038] env[62183]: DEBUG nova.compute.manager [req-1d306f7b-acff-4f6c-bfdd-9902c147a5ab req-0b8cf4ba-587d-49a3-ab46-58a427dfde5d service nova] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Refreshing instance network info cache due to event network-changed-48058051-c983-4135-8170-3395eeadc4e3. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 614.124038] env[62183]: DEBUG oslo_concurrency.lockutils [req-1d306f7b-acff-4f6c-bfdd-9902c147a5ab req-0b8cf4ba-587d-49a3-ab46-58a427dfde5d service nova] Acquiring lock "refresh_cache-79c10e4e-98ab-4259-8ff5-72bfa39860be" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.124038] env[62183]: DEBUG oslo_concurrency.lockutils [req-1d306f7b-acff-4f6c-bfdd-9902c147a5ab req-0b8cf4ba-587d-49a3-ab46-58a427dfde5d service nova] Acquired lock "refresh_cache-79c10e4e-98ab-4259-8ff5-72bfa39860be" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.124038] env[62183]: DEBUG nova.network.neutron [req-1d306f7b-acff-4f6c-bfdd-9902c147a5ab req-0b8cf4ba-587d-49a3-ab46-58a427dfde5d service nova] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Refreshing network info cache for port 48058051-c983-4135-8170-3395eeadc4e3 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 614.136272] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528f1add-e9b0-5bd5-c82a-92062b345bc4, 'name': SearchDatastore_Task, 'duration_secs': 0.010192} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.136757] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.137026] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 614.137459] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0adb94e8-c08b-4657-b545-fbe1eeefd227 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.140468] env[62183]: DEBUG nova.network.neutron [-] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.151501] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 614.151501] env[62183]: value = "task-1386900" [ 614.151501] env[62183]: _type = "Task" [ 614.151501] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.164264] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.262017] env[62183]: INFO nova.compute.manager [-] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Took 1.03 seconds to deallocate network for instance. [ 614.538386] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 614.543335] env[62183]: DEBUG nova.scheduler.client.report [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 614.592727] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 614.592727] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 614.592864] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.592993] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 614.593197] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.593358] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 614.593579] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 614.593750] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 614.593914] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 614.594085] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 614.594254] env[62183]: DEBUG nova.virt.hardware [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.595557] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4fef97b-896d-4dc1-8f7d-7c8a72b90906 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.607604] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc282770-31f8-4889-9d3a-c34186b2d248 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.645108] env[62183]: DEBUG nova.network.neutron [-] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.659027] env[62183]: DEBUG nova.network.neutron [req-1d306f7b-acff-4f6c-bfdd-9902c147a5ab req-0b8cf4ba-587d-49a3-ab46-58a427dfde5d service nova] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.666092] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386900, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.758020] env[62183]: DEBUG nova.network.neutron [req-1d306f7b-acff-4f6c-bfdd-9902c147a5ab req-0b8cf4ba-587d-49a3-ab46-58a427dfde5d service nova] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.839107] env[62183]: INFO nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Took 0.58 seconds to detach 1 volumes for instance. [ 614.843448] env[62183]: DEBUG nova.compute.claims [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.843528] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.924688] env[62183]: ERROR nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0b1c71e3-677d-4088-977d-208deb7610fe, please check neutron logs for more information. [ 614.924688] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.924688] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.924688] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.924688] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.924688] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.924688] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.924688] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.924688] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.924688] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 614.924688] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.924688] env[62183]: ERROR nova.compute.manager raise self.value [ 614.924688] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.924688] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.924688] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.924688] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.925407] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.925407] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.925407] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0b1c71e3-677d-4088-977d-208deb7610fe, please check neutron logs for more information. [ 614.925407] env[62183]: ERROR nova.compute.manager [ 614.925407] env[62183]: Traceback (most recent call last): [ 614.925407] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.925407] env[62183]: listener.cb(fileno) [ 614.925407] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.925407] env[62183]: result = function(*args, **kwargs) [ 614.925407] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.925407] env[62183]: return func(*args, **kwargs) [ 614.925407] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.925407] env[62183]: raise e [ 614.925407] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.925407] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 614.925407] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.925407] env[62183]: created_port_ids = self._update_ports_for_instance( [ 614.925407] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.925407] env[62183]: with excutils.save_and_reraise_exception(): [ 614.925407] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.925407] env[62183]: self.force_reraise() [ 614.925407] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.925407] env[62183]: raise self.value [ 614.925407] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.925407] env[62183]: updated_port = self._update_port( [ 614.925407] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.925407] env[62183]: _ensure_no_port_binding_failure(port) [ 614.925407] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.925407] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.926556] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 0b1c71e3-677d-4088-977d-208deb7610fe, please check neutron logs for more information. [ 614.926556] env[62183]: Removing descriptor: 14 [ 614.926556] env[62183]: ERROR nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0b1c71e3-677d-4088-977d-208deb7610fe, please check neutron logs for more information. [ 614.926556] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Traceback (most recent call last): [ 614.926556] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.926556] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] yield resources [ 614.926556] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.926556] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self.driver.spawn(context, instance, image_meta, [ 614.926556] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 614.926556] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.926556] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.926556] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] vm_ref = self.build_virtual_machine(instance, [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] for vif in network_info: [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] return self._sync_wrapper(fn, *args, **kwargs) [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self.wait() [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self[:] = self._gt.wait() [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] return self._exit_event.wait() [ 614.927179] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] result = hub.switch() [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] return self.greenlet.switch() [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] result = function(*args, **kwargs) [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] return func(*args, **kwargs) [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] raise e [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] nwinfo = self.network_api.allocate_for_instance( [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.927745] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] created_port_ids = self._update_ports_for_instance( [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] with excutils.save_and_reraise_exception(): [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self.force_reraise() [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] raise self.value [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] updated_port = self._update_port( [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] _ensure_no_port_binding_failure(port) [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.928270] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] raise exception.PortBindingFailed(port_id=port['id']) [ 614.929312] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] nova.exception.PortBindingFailed: Binding failed for port 0b1c71e3-677d-4088-977d-208deb7610fe, please check neutron logs for more information. [ 614.929312] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] [ 614.929312] env[62183]: INFO nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Terminating instance [ 614.929312] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Acquiring lock "refresh_cache-ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.929312] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Acquired lock "refresh_cache-ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.929312] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 615.063303] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.063303] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 615.066382] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.273s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.150323] env[62183]: INFO nova.compute.manager [-] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Took 1.03 seconds to deallocate network for instance. [ 615.151654] env[62183]: DEBUG nova.compute.claims [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 615.152474] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.162693] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386900, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.264193] env[62183]: DEBUG oslo_concurrency.lockutils [req-1d306f7b-acff-4f6c-bfdd-9902c147a5ab req-0b8cf4ba-587d-49a3-ab46-58a427dfde5d service nova] Releasing lock "refresh_cache-79c10e4e-98ab-4259-8ff5-72bfa39860be" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.264193] env[62183]: DEBUG nova.compute.manager [req-1d306f7b-acff-4f6c-bfdd-9902c147a5ab req-0b8cf4ba-587d-49a3-ab46-58a427dfde5d service nova] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Received event network-vif-deleted-48058051-c983-4135-8170-3395eeadc4e3 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.452209] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.557067] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.572658] env[62183]: DEBUG nova.compute.utils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.574119] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.574333] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 615.626127] env[62183]: DEBUG nova.policy [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7deeb5ccaca42e29f0cb145a17e1097', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37d4936e2dc34e9d9ed73455081ee4ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 615.662999] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386900, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.442205} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.665547] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 615.665767] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 615.666208] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6889f546-3169-4744-9aa0-94b3b802c409 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.672190] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 615.672190] env[62183]: value = "task-1386901" [ 615.672190] env[62183]: _type = "Task" [ 615.672190] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.681608] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386901, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.876555] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Acquiring lock "b0761551-b2a3-4a38-a32d-f94a36c47d94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.876783] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Lock "b0761551-b2a3-4a38-a32d-f94a36c47d94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.944248] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Successfully created port: 82ae4109-324e-4069-adad-1060e2f15a09 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.028718] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19186a1d-663a-435a-9e40-679e5a9adaa1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.041536] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc22d19-52b3-46ea-81f6-3e2c3fef900f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.073889] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Releasing lock "refresh_cache-ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.074370] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.074625] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 616.075423] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 616.078203] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-19a4a533-e119-445e-9dad-968dbd3e77b7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.082514] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664c47ac-0baa-4576-8d78-bad94db34c70 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.096039] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebd7e55-91f1-46f1-bd34-1c41774b01e5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.100892] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b269d27-9461-422f-bbe7-6d1db8a13e27 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.120123] env[62183]: DEBUG nova.compute.provider_tree [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.134615] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65 could not be found. [ 616.135983] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 616.135983] env[62183]: INFO nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Took 0.06 seconds to destroy the instance on the hypervisor. [ 616.135983] env[62183]: DEBUG oslo.service.loopingcall [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.136151] env[62183]: DEBUG nova.compute.manager [-] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.136302] env[62183]: DEBUG nova.network.neutron [-] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 616.161531] env[62183]: DEBUG nova.network.neutron [-] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.164414] env[62183]: DEBUG nova.compute.manager [req-186d196b-08bf-4aa6-b12d-be0ab4bea564 req-c2c715de-e07b-4348-b2a7-680b2aedcfcb service nova] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Received event network-changed-0b1c71e3-677d-4088-977d-208deb7610fe {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.164617] env[62183]: DEBUG nova.compute.manager [req-186d196b-08bf-4aa6-b12d-be0ab4bea564 req-c2c715de-e07b-4348-b2a7-680b2aedcfcb service nova] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Refreshing instance network info cache due to event network-changed-0b1c71e3-677d-4088-977d-208deb7610fe. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 616.164826] env[62183]: DEBUG oslo_concurrency.lockutils [req-186d196b-08bf-4aa6-b12d-be0ab4bea564 req-c2c715de-e07b-4348-b2a7-680b2aedcfcb service nova] Acquiring lock "refresh_cache-ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.164975] env[62183]: DEBUG oslo_concurrency.lockutils [req-186d196b-08bf-4aa6-b12d-be0ab4bea564 req-c2c715de-e07b-4348-b2a7-680b2aedcfcb service nova] Acquired lock "refresh_cache-ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.165159] env[62183]: DEBUG nova.network.neutron [req-186d196b-08bf-4aa6-b12d-be0ab4bea564 req-c2c715de-e07b-4348-b2a7-680b2aedcfcb service nova] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Refreshing network info cache for port 0b1c71e3-677d-4088-977d-208deb7610fe {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 616.183072] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386901, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064488} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.183358] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 616.184786] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36525fb-115e-43e4-a965-6431c686cfe6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.204528] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Reconfiguring VM instance instance-00000011 to attach disk [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 616.206115] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0f72d19-8c56-40bf-9fea-371394e5dd75 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.223884] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Acquiring lock "430dcd52-17ce-4710-b9c8-39c28410e356" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.224139] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Lock "430dcd52-17ce-4710-b9c8-39c28410e356" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.230580] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 616.230580] env[62183]: value = "task-1386902" [ 616.230580] env[62183]: _type = "Task" [ 616.230580] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.242904] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386902, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.623171] env[62183]: DEBUG nova.scheduler.client.report [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 616.667298] env[62183]: DEBUG nova.network.neutron [-] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.687041] env[62183]: DEBUG nova.network.neutron [req-186d196b-08bf-4aa6-b12d-be0ab4bea564 req-c2c715de-e07b-4348-b2a7-680b2aedcfcb service nova] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.740882] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386902, 'name': ReconfigVM_Task, 'duration_secs': 0.309208} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.741142] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Reconfigured VM instance instance-00000011 to attach disk [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1/a43e24b0-c0f1-4e02-894d-0d2c8ed884d1.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 616.741758] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c655065-c46e-4b37-a0c4-b47288b1d67a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.748343] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 616.748343] env[62183]: value = "task-1386903" [ 616.748343] env[62183]: _type = "Task" [ 616.748343] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.758398] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386903, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.836155] env[62183]: DEBUG nova.network.neutron [req-186d196b-08bf-4aa6-b12d-be0ab4bea564 req-c2c715de-e07b-4348-b2a7-680b2aedcfcb service nova] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.092487] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 617.121501] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 617.121762] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 617.121923] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.122230] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 617.122539] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.122843] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 617.123251] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 617.123654] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 617.124226] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 617.124549] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 617.124897] env[62183]: DEBUG nova.virt.hardware [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.126617] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f590eda-e936-4b9a-9fb3-922649311fee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.130371] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.065s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.130965] env[62183]: ERROR nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d4e93cda-fa04-4fb5-8636-85f18a5771c9, please check neutron logs for more information. [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Traceback (most recent call last): [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self.driver.spawn(context, instance, image_meta, [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] vm_ref = self.build_virtual_machine(instance, [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.130965] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] for vif in network_info: [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] return self._sync_wrapper(fn, *args, **kwargs) [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self.wait() [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self[:] = self._gt.wait() [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] return self._exit_event.wait() [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] current.throw(*self._exc) [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.131510] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] result = function(*args, **kwargs) [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] return func(*args, **kwargs) [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] raise e [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] nwinfo = self.network_api.allocate_for_instance( [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] created_port_ids = self._update_ports_for_instance( [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] with excutils.save_and_reraise_exception(): [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] self.force_reraise() [ 617.132083] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.132693] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] raise self.value [ 617.132693] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.132693] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] updated_port = self._update_port( [ 617.132693] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.132693] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] _ensure_no_port_binding_failure(port) [ 617.132693] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.132693] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] raise exception.PortBindingFailed(port_id=port['id']) [ 617.132693] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] nova.exception.PortBindingFailed: Binding failed for port d4e93cda-fa04-4fb5-8636-85f18a5771c9, please check neutron logs for more information. [ 617.132693] env[62183]: ERROR nova.compute.manager [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] [ 617.132693] env[62183]: DEBUG nova.compute.utils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Binding failed for port d4e93cda-fa04-4fb5-8636-85f18a5771c9, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 617.134612] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.867s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.138615] env[62183]: INFO nova.compute.claims [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.143675] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Build of instance 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe was re-scheduled: Binding failed for port d4e93cda-fa04-4fb5-8636-85f18a5771c9, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 617.144168] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 617.144395] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Acquiring lock "refresh_cache-1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.144544] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Acquired lock "refresh_cache-1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.144697] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.152964] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974d267c-2f3d-46bd-820f-d113508f12c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.171915] env[62183]: INFO nova.compute.manager [-] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Took 1.04 seconds to deallocate network for instance. [ 617.175014] env[62183]: DEBUG nova.compute.claims [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 617.175199] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.258394] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386903, 'name': Rename_Task, 'duration_secs': 0.131722} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.258669] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 617.258913] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f696c32-283d-4777-b4b0-c8eb651c42d5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.265551] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Waiting for the task: (returnval){ [ 617.265551] env[62183]: value = "task-1386904" [ 617.265551] env[62183]: _type = "Task" [ 617.265551] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.269488] env[62183]: ERROR nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 82ae4109-324e-4069-adad-1060e2f15a09, please check neutron logs for more information. [ 617.269488] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.269488] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.269488] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.269488] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.269488] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.269488] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.269488] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.269488] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.269488] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 617.269488] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.269488] env[62183]: ERROR nova.compute.manager raise self.value [ 617.269488] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.269488] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.269488] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.269488] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.269987] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.269987] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.269987] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 82ae4109-324e-4069-adad-1060e2f15a09, please check neutron logs for more information. [ 617.269987] env[62183]: ERROR nova.compute.manager [ 617.269987] env[62183]: Traceback (most recent call last): [ 617.269987] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.269987] env[62183]: listener.cb(fileno) [ 617.269987] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.269987] env[62183]: result = function(*args, **kwargs) [ 617.269987] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 617.269987] env[62183]: return func(*args, **kwargs) [ 617.269987] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.269987] env[62183]: raise e [ 617.269987] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.269987] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 617.269987] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.269987] env[62183]: created_port_ids = self._update_ports_for_instance( [ 617.269987] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.269987] env[62183]: with excutils.save_and_reraise_exception(): [ 617.269987] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.269987] env[62183]: self.force_reraise() [ 617.269987] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.269987] env[62183]: raise self.value [ 617.269987] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.269987] env[62183]: updated_port = self._update_port( [ 617.269987] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.269987] env[62183]: _ensure_no_port_binding_failure(port) [ 617.269987] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.269987] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.270847] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 82ae4109-324e-4069-adad-1060e2f15a09, please check neutron logs for more information. [ 617.270847] env[62183]: Removing descriptor: 14 [ 617.270847] env[62183]: ERROR nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 82ae4109-324e-4069-adad-1060e2f15a09, please check neutron logs for more information. [ 617.270847] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Traceback (most recent call last): [ 617.270847] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 617.270847] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] yield resources [ 617.270847] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.270847] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self.driver.spawn(context, instance, image_meta, [ 617.270847] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 617.270847] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.270847] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.270847] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] vm_ref = self.build_virtual_machine(instance, [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] for vif in network_info: [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] return self._sync_wrapper(fn, *args, **kwargs) [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self.wait() [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self[:] = self._gt.wait() [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] return self._exit_event.wait() [ 617.271214] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] result = hub.switch() [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] return self.greenlet.switch() [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] result = function(*args, **kwargs) [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] return func(*args, **kwargs) [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] raise e [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] nwinfo = self.network_api.allocate_for_instance( [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.271611] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] created_port_ids = self._update_ports_for_instance( [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] with excutils.save_and_reraise_exception(): [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self.force_reraise() [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] raise self.value [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] updated_port = self._update_port( [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] _ensure_no_port_binding_failure(port) [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.271988] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] raise exception.PortBindingFailed(port_id=port['id']) [ 617.272534] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] nova.exception.PortBindingFailed: Binding failed for port 82ae4109-324e-4069-adad-1060e2f15a09, please check neutron logs for more information. [ 617.272534] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] [ 617.272534] env[62183]: INFO nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Terminating instance [ 617.273020] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "refresh_cache-bed9c66e-a52f-4ec9-a190-a46c92547c09" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.273181] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "refresh_cache-bed9c66e-a52f-4ec9-a190-a46c92547c09" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.273346] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.278566] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.339221] env[62183]: DEBUG oslo_concurrency.lockutils [req-186d196b-08bf-4aa6-b12d-be0ab4bea564 req-c2c715de-e07b-4348-b2a7-680b2aedcfcb service nova] Releasing lock "refresh_cache-ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.339444] env[62183]: DEBUG nova.compute.manager [req-186d196b-08bf-4aa6-b12d-be0ab4bea564 req-c2c715de-e07b-4348-b2a7-680b2aedcfcb service nova] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Received event network-vif-deleted-0b1c71e3-677d-4088-977d-208deb7610fe {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.675970] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.762382] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.782922] env[62183]: DEBUG oslo_vmware.api [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Task: {'id': task-1386904, 'name': PowerOnVM_Task, 'duration_secs': 0.444949} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.783809] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 617.784041] env[62183]: DEBUG nova.compute.manager [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 617.786112] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d69a0a-4057-4d95-835d-3f60eda7bfca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.797538] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.883181] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.201406] env[62183]: DEBUG nova.compute.manager [req-e0e314dd-a9bf-4f18-b27c-8c69792ff04a req-5b9488fd-fa62-4e90-990a-d4220ee71cc9 service nova] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Received event network-changed-82ae4109-324e-4069-adad-1060e2f15a09 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.201828] env[62183]: DEBUG nova.compute.manager [req-e0e314dd-a9bf-4f18-b27c-8c69792ff04a req-5b9488fd-fa62-4e90-990a-d4220ee71cc9 service nova] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Refreshing instance network info cache due to event network-changed-82ae4109-324e-4069-adad-1060e2f15a09. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.201828] env[62183]: DEBUG oslo_concurrency.lockutils [req-e0e314dd-a9bf-4f18-b27c-8c69792ff04a req-5b9488fd-fa62-4e90-990a-d4220ee71cc9 service nova] Acquiring lock "refresh_cache-bed9c66e-a52f-4ec9-a190-a46c92547c09" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.269666] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Releasing lock "refresh_cache-1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.269853] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 618.270041] env[62183]: DEBUG nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.270272] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.293306] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.306149] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.387019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "refresh_cache-bed9c66e-a52f-4ec9-a190-a46c92547c09" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.387019] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 618.387152] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 618.387856] env[62183]: DEBUG oslo_concurrency.lockutils [req-e0e314dd-a9bf-4f18-b27c-8c69792ff04a req-5b9488fd-fa62-4e90-990a-d4220ee71cc9 service nova] Acquired lock "refresh_cache-bed9c66e-a52f-4ec9-a190-a46c92547c09" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.388068] env[62183]: DEBUG nova.network.neutron [req-e0e314dd-a9bf-4f18-b27c-8c69792ff04a req-5b9488fd-fa62-4e90-990a-d4220ee71cc9 service nova] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Refreshing network info cache for port 82ae4109-324e-4069-adad-1060e2f15a09 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.389343] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83f76d1b-86e5-4303-92de-f1dcd0edec5f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.401519] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d0b111-ac15-499c-8f76-9f6111570632 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.426244] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bed9c66e-a52f-4ec9-a190-a46c92547c09 could not be found. [ 618.426478] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 618.426660] env[62183]: INFO nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Took 0.04 seconds to destroy the instance on the hypervisor. [ 618.426901] env[62183]: DEBUG oslo.service.loopingcall [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.429711] env[62183]: DEBUG nova.compute.manager [-] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.429834] env[62183]: DEBUG nova.network.neutron [-] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.449492] env[62183]: DEBUG nova.network.neutron [-] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.560028] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f3b47b-af66-41b9-b2c5-9275fed91681 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.567399] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d147b09-b5db-430b-8b45-525bdd9f567f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.604253] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34120ea-f5c9-4f29-87b3-8650f597e7f8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.614533] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36f22ad-8714-4cda-abe0-614209b5c820 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.634886] env[62183]: DEBUG nova.compute.provider_tree [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.802193] env[62183]: DEBUG nova.network.neutron [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.854418] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "a43e24b0-c0f1-4e02-894d-0d2c8ed884d1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.854681] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "a43e24b0-c0f1-4e02-894d-0d2c8ed884d1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.854881] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "a43e24b0-c0f1-4e02-894d-0d2c8ed884d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.855073] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "a43e24b0-c0f1-4e02-894d-0d2c8ed884d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.855268] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "a43e24b0-c0f1-4e02-894d-0d2c8ed884d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.857391] env[62183]: INFO nova.compute.manager [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Terminating instance [ 618.859089] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "refresh_cache-a43e24b0-c0f1-4e02-894d-0d2c8ed884d1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.859248] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquired lock "refresh_cache-a43e24b0-c0f1-4e02-894d-0d2c8ed884d1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.859424] env[62183]: DEBUG nova.network.neutron [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 618.907303] env[62183]: DEBUG nova.network.neutron [req-e0e314dd-a9bf-4f18-b27c-8c69792ff04a req-5b9488fd-fa62-4e90-990a-d4220ee71cc9 service nova] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.951797] env[62183]: DEBUG nova.network.neutron [-] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.990688] env[62183]: DEBUG nova.network.neutron [req-e0e314dd-a9bf-4f18-b27c-8c69792ff04a req-5b9488fd-fa62-4e90-990a-d4220ee71cc9 service nova] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.139069] env[62183]: DEBUG nova.scheduler.client.report [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 619.304951] env[62183]: INFO nova.compute.manager [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] [instance: 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe] Took 1.03 seconds to deallocate network for instance. [ 619.376253] env[62183]: DEBUG nova.network.neutron [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.428291] env[62183]: DEBUG nova.network.neutron [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.453816] env[62183]: INFO nova.compute.manager [-] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Took 1.02 seconds to deallocate network for instance. [ 619.457535] env[62183]: DEBUG nova.compute.claims [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 619.457535] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.494937] env[62183]: DEBUG oslo_concurrency.lockutils [req-e0e314dd-a9bf-4f18-b27c-8c69792ff04a req-5b9488fd-fa62-4e90-990a-d4220ee71cc9 service nova] Releasing lock "refresh_cache-bed9c66e-a52f-4ec9-a190-a46c92547c09" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.494937] env[62183]: DEBUG nova.compute.manager [req-e0e314dd-a9bf-4f18-b27c-8c69792ff04a req-5b9488fd-fa62-4e90-990a-d4220ee71cc9 service nova] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Received event network-vif-deleted-82ae4109-324e-4069-adad-1060e2f15a09 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.642990] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.643658] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.646234] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.477s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.647598] env[62183]: INFO nova.compute.claims [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.930579] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Releasing lock "refresh_cache-a43e24b0-c0f1-4e02-894d-0d2c8ed884d1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.931036] env[62183]: DEBUG nova.compute.manager [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 619.931341] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.932145] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a86750-f132-400a-965d-1918b44d5345 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.941092] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 619.941368] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6ed4043-1444-4912-adb1-4e0baa3ef507 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.948358] env[62183]: DEBUG oslo_vmware.api [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 619.948358] env[62183]: value = "task-1386905" [ 619.948358] env[62183]: _type = "Task" [ 619.948358] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.955279] env[62183]: DEBUG oslo_vmware.api [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386905, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.152008] env[62183]: DEBUG nova.compute.utils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 620.155918] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 620.155918] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 620.202464] env[62183]: DEBUG nova.policy [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0feda18d16da4f83a96d0e684d84975b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa507ce81ee94145a6f39bd7e7046abc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 620.334859] env[62183]: INFO nova.scheduler.client.report [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Deleted allocations for instance 1c4a3849-45ea-4c33-8b1f-f52ea5fefafe [ 620.459753] env[62183]: DEBUG oslo_vmware.api [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386905, 'name': PowerOffVM_Task, 'duration_secs': 0.110484} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.460177] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 620.462404] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 620.462404] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-531e2c52-e4fa-444e-89eb-880e141b07cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.487250] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 620.487250] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 620.487250] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Deleting the datastore file [datastore1] a43e24b0-c0f1-4e02-894d-0d2c8ed884d1 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 620.487250] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2bc3425-f9ce-4780-9355-76cab7629c8a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.492797] env[62183]: DEBUG oslo_vmware.api [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for the task: (returnval){ [ 620.492797] env[62183]: value = "task-1386907" [ 620.492797] env[62183]: _type = "Task" [ 620.492797] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.501489] env[62183]: DEBUG oslo_vmware.api [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386907, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.502413] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Successfully created port: 1655cefc-8ccb-4fa0-adcf-d4a850b8f691 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.656690] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.845481] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2afc8c60-facc-47c8-893b-209032b488ad tempest-ImagesNegativeTestJSON-1875692687 tempest-ImagesNegativeTestJSON-1875692687-project-member] Lock "1c4a3849-45ea-4c33-8b1f-f52ea5fefafe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.368s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.007397] env[62183]: DEBUG oslo_vmware.api [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Task: {'id': task-1386907, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097534} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.009898] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 621.010017] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 621.010216] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.010417] env[62183]: INFO nova.compute.manager [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Took 1.08 seconds to destroy the instance on the hypervisor. [ 621.010658] env[62183]: DEBUG oslo.service.loopingcall [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.011074] env[62183]: DEBUG nova.compute.manager [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.011178] env[62183]: DEBUG nova.network.neutron [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.028896] env[62183]: DEBUG nova.network.neutron [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.080480] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755f9d39-bec1-45b2-aee0-8a744cf31c59 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.088677] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d49702e-7d14-4c58-84c9-8839556bb87b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.128519] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daeea679-1763-4815-a537-66649fc5b239 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.136225] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1069101a-05c6-4407-9cfc-549cdbfd89ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.150848] env[62183]: DEBUG nova.compute.provider_tree [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.348715] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.369953] env[62183]: DEBUG nova.compute.manager [req-c5278880-b02f-4d7a-9915-ecfe8113895e req-4282cb07-b75b-462a-a2a3-58839d81f85b service nova] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Received event network-changed-1655cefc-8ccb-4fa0-adcf-d4a850b8f691 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 621.370176] env[62183]: DEBUG nova.compute.manager [req-c5278880-b02f-4d7a-9915-ecfe8113895e req-4282cb07-b75b-462a-a2a3-58839d81f85b service nova] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Refreshing instance network info cache due to event network-changed-1655cefc-8ccb-4fa0-adcf-d4a850b8f691. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 621.370458] env[62183]: DEBUG oslo_concurrency.lockutils [req-c5278880-b02f-4d7a-9915-ecfe8113895e req-4282cb07-b75b-462a-a2a3-58839d81f85b service nova] Acquiring lock "refresh_cache-650b8016-5a30-405a-b85c-6153228c9603" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.370702] env[62183]: DEBUG oslo_concurrency.lockutils [req-c5278880-b02f-4d7a-9915-ecfe8113895e req-4282cb07-b75b-462a-a2a3-58839d81f85b service nova] Acquired lock "refresh_cache-650b8016-5a30-405a-b85c-6153228c9603" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.370772] env[62183]: DEBUG nova.network.neutron [req-c5278880-b02f-4d7a-9915-ecfe8113895e req-4282cb07-b75b-462a-a2a3-58839d81f85b service nova] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Refreshing network info cache for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 621.531732] env[62183]: DEBUG nova.network.neutron [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.593826] env[62183]: ERROR nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691, please check neutron logs for more information. [ 621.593826] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 621.593826] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.593826] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 621.593826] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.593826] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 621.593826] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.593826] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 621.593826] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.593826] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 621.593826] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.593826] env[62183]: ERROR nova.compute.manager raise self.value [ 621.593826] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.593826] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 621.593826] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.593826] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 621.594373] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.594373] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 621.594373] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691, please check neutron logs for more information. [ 621.594373] env[62183]: ERROR nova.compute.manager [ 621.594373] env[62183]: Traceback (most recent call last): [ 621.594373] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 621.594373] env[62183]: listener.cb(fileno) [ 621.594373] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.594373] env[62183]: result = function(*args, **kwargs) [ 621.594373] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.594373] env[62183]: return func(*args, **kwargs) [ 621.594373] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.594373] env[62183]: raise e [ 621.594373] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.594373] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 621.594373] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.594373] env[62183]: created_port_ids = self._update_ports_for_instance( [ 621.594373] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.594373] env[62183]: with excutils.save_and_reraise_exception(): [ 621.594373] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.594373] env[62183]: self.force_reraise() [ 621.594373] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.594373] env[62183]: raise self.value [ 621.594373] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.594373] env[62183]: updated_port = self._update_port( [ 621.594373] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.594373] env[62183]: _ensure_no_port_binding_failure(port) [ 621.594373] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.594373] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 621.595208] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691, please check neutron logs for more information. [ 621.595208] env[62183]: Removing descriptor: 14 [ 621.654193] env[62183]: DEBUG nova.scheduler.client.report [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 621.668037] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.692642] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:06:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='8c0584e2-43a0-4379-935a-aa5f18360265',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1489051059',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.692883] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.693051] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.693242] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.693388] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.693533] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.693737] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.694112] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.694112] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.694280] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.694448] env[62183]: DEBUG nova.virt.hardware [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.695303] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b9b480-a3ea-4b04-9e97-63e9e116167f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.703639] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c34dd7-4a1e-4925-b951-f9d52db10bd7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.717077] env[62183]: ERROR nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691, please check neutron logs for more information. [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] Traceback (most recent call last): [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] yield resources [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self.driver.spawn(context, instance, image_meta, [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] vm_ref = self.build_virtual_machine(instance, [ 621.717077] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] for vif in network_info: [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] return self._sync_wrapper(fn, *args, **kwargs) [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self.wait() [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self[:] = self._gt.wait() [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] return self._exit_event.wait() [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 621.717464] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] current.throw(*self._exc) [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] result = function(*args, **kwargs) [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] return func(*args, **kwargs) [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] raise e [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] nwinfo = self.network_api.allocate_for_instance( [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] created_port_ids = self._update_ports_for_instance( [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] with excutils.save_and_reraise_exception(): [ 621.717801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self.force_reraise() [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] raise self.value [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] updated_port = self._update_port( [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] _ensure_no_port_binding_failure(port) [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] raise exception.PortBindingFailed(port_id=port['id']) [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] nova.exception.PortBindingFailed: Binding failed for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691, please check neutron logs for more information. [ 621.718198] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] [ 621.718198] env[62183]: INFO nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Terminating instance [ 621.719374] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "refresh_cache-650b8016-5a30-405a-b85c-6153228c9603" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.874108] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.895172] env[62183]: DEBUG nova.network.neutron [req-c5278880-b02f-4d7a-9915-ecfe8113895e req-4282cb07-b75b-462a-a2a3-58839d81f85b service nova] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.985361] env[62183]: DEBUG nova.network.neutron [req-c5278880-b02f-4d7a-9915-ecfe8113895e req-4282cb07-b75b-462a-a2a3-58839d81f85b service nova] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.034153] env[62183]: INFO nova.compute.manager [-] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Took 1.02 seconds to deallocate network for instance. [ 622.159679] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.160353] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 622.163049] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.329s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.167021] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.167021] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 622.167021] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.841s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.167621] env[62183]: INFO nova.compute.claims [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.171882] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f91fd7-998c-49df-a611-af438b1f7a72 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.180996] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dac52d3-cb5f-4406-b7c6-ddab86748e8a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.205534] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35669a87-5aa6-47ad-9d2a-406a3a99c1dc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.213609] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7b081d-8f9c-461d-a9fc-128b3c9f2677 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.245474] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181369MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 622.245643] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.490335] env[62183]: DEBUG oslo_concurrency.lockutils [req-c5278880-b02f-4d7a-9915-ecfe8113895e req-4282cb07-b75b-462a-a2a3-58839d81f85b service nova] Releasing lock "refresh_cache-650b8016-5a30-405a-b85c-6153228c9603" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.490335] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquired lock "refresh_cache-650b8016-5a30-405a-b85c-6153228c9603" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.490335] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 622.541536] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.674868] env[62183]: DEBUG nova.compute.utils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.679407] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.679407] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 622.749506] env[62183]: DEBUG nova.policy [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f387516e32e24f5d872e72bd30bb59f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f6e78720aec4493afdd15ca78aac5f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 623.028314] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.122563] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.177780] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 623.262661] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Successfully created port: 60361efe-6879-402d-aa00-16d78c27baac {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.496590] env[62183]: DEBUG nova.compute.manager [req-f69908eb-b1ce-4df2-af55-5301d798ad40 req-70d28c22-f800-4be4-bb70-4e17386c3d47 service nova] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Received event network-vif-deleted-1655cefc-8ccb-4fa0-adcf-d4a850b8f691 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.625045] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Releasing lock "refresh_cache-650b8016-5a30-405a-b85c-6153228c9603" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.625451] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 623.625644] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 623.626256] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a77dd5c-713e-4dc0-9901-7969af584c04 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.635407] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2bb5b22-8ebd-47b8-a3ba-e73898964889 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.658986] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 650b8016-5a30-405a-b85c-6153228c9603 could not be found. [ 623.659227] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 623.659409] env[62183]: INFO nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Took 0.03 seconds to destroy the instance on the hypervisor. [ 623.659650] env[62183]: DEBUG oslo.service.loopingcall [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.662117] env[62183]: DEBUG nova.compute.manager [-] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.662223] env[62183]: DEBUG nova.network.neutron [-] [instance: 650b8016-5a30-405a-b85c-6153228c9603] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.674627] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7b1b48-295f-442f-b503-b94d0634cac8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.681688] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a228288-de9e-4c6f-b413-44e62f3ee44f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.686711] env[62183]: DEBUG nova.network.neutron [-] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.719738] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72144a1-989f-44dd-b264-bd708a8967d3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.729659] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db06c9e-fdcb-4153-bd8c-8535db1518ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.743565] env[62183]: DEBUG nova.compute.provider_tree [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.193736] env[62183]: DEBUG nova.network.neutron [-] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.195842] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 624.237078] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.237350] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.237531] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.237716] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.237860] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.238009] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.238613] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.238819] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.239013] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.239185] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.239358] env[62183]: DEBUG nova.virt.hardware [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.240311] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b48497-2fb7-4200-b2c6-5fe5bb80fa80 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.248621] env[62183]: DEBUG nova.scheduler.client.report [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 624.253512] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7363bd-0fc3-49ed-875a-bd6ceb518b7d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.488120] env[62183]: ERROR nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 60361efe-6879-402d-aa00-16d78c27baac, please check neutron logs for more information. [ 624.488120] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.488120] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.488120] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.488120] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.488120] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.488120] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.488120] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.488120] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.488120] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 624.488120] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.488120] env[62183]: ERROR nova.compute.manager raise self.value [ 624.488120] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.488120] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.488120] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.488120] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.488536] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.488536] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.488536] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 60361efe-6879-402d-aa00-16d78c27baac, please check neutron logs for more information. [ 624.488536] env[62183]: ERROR nova.compute.manager [ 624.488536] env[62183]: Traceback (most recent call last): [ 624.488536] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.488536] env[62183]: listener.cb(fileno) [ 624.488536] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.488536] env[62183]: result = function(*args, **kwargs) [ 624.488536] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.488536] env[62183]: return func(*args, **kwargs) [ 624.488536] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.488536] env[62183]: raise e [ 624.488536] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.488536] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 624.488536] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.488536] env[62183]: created_port_ids = self._update_ports_for_instance( [ 624.488536] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.488536] env[62183]: with excutils.save_and_reraise_exception(): [ 624.488536] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.488536] env[62183]: self.force_reraise() [ 624.488536] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.488536] env[62183]: raise self.value [ 624.488536] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.488536] env[62183]: updated_port = self._update_port( [ 624.488536] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.488536] env[62183]: _ensure_no_port_binding_failure(port) [ 624.488536] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.488536] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.489246] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 60361efe-6879-402d-aa00-16d78c27baac, please check neutron logs for more information. [ 624.489246] env[62183]: Removing descriptor: 14 [ 624.489246] env[62183]: ERROR nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 60361efe-6879-402d-aa00-16d78c27baac, please check neutron logs for more information. [ 624.489246] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Traceback (most recent call last): [ 624.489246] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 624.489246] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] yield resources [ 624.489246] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.489246] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self.driver.spawn(context, instance, image_meta, [ 624.489246] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 624.489246] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.489246] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.489246] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] vm_ref = self.build_virtual_machine(instance, [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] for vif in network_info: [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] return self._sync_wrapper(fn, *args, **kwargs) [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self.wait() [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self[:] = self._gt.wait() [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] return self._exit_event.wait() [ 624.489563] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] result = hub.switch() [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] return self.greenlet.switch() [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] result = function(*args, **kwargs) [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] return func(*args, **kwargs) [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] raise e [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] nwinfo = self.network_api.allocate_for_instance( [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 624.489914] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] created_port_ids = self._update_ports_for_instance( [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] with excutils.save_and_reraise_exception(): [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self.force_reraise() [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] raise self.value [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] updated_port = self._update_port( [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] _ensure_no_port_binding_failure(port) [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.490289] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] raise exception.PortBindingFailed(port_id=port['id']) [ 624.490655] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] nova.exception.PortBindingFailed: Binding failed for port 60361efe-6879-402d-aa00-16d78c27baac, please check neutron logs for more information. [ 624.490655] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] [ 624.490655] env[62183]: INFO nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Terminating instance [ 624.490988] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "refresh_cache-cede536f-eac5-4dd0-b4b5-74cfbd24300b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.491158] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquired lock "refresh_cache-cede536f-eac5-4dd0-b4b5-74cfbd24300b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.491317] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.699372] env[62183]: INFO nova.compute.manager [-] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Took 1.04 seconds to deallocate network for instance. [ 624.702157] env[62183]: DEBUG nova.compute.claims [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 624.702342] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.753416] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.753931] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 624.756609] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.517s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.757295] env[62183]: DEBUG nova.objects.instance [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62183) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 625.007199] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.056907] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.261667] env[62183]: DEBUG nova.compute.utils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.265672] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.265891] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 625.339044] env[62183]: DEBUG nova.policy [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f387516e32e24f5d872e72bd30bb59f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f6e78720aec4493afdd15ca78aac5f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.524574] env[62183]: DEBUG nova.compute.manager [req-aa73e81e-d653-495a-af20-a9f7092c99cc req-a68af173-b106-4e58-9a90-acb4b8a8d4ef service nova] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Received event network-changed-60361efe-6879-402d-aa00-16d78c27baac {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.524773] env[62183]: DEBUG nova.compute.manager [req-aa73e81e-d653-495a-af20-a9f7092c99cc req-a68af173-b106-4e58-9a90-acb4b8a8d4ef service nova] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Refreshing instance network info cache due to event network-changed-60361efe-6879-402d-aa00-16d78c27baac. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 625.524961] env[62183]: DEBUG oslo_concurrency.lockutils [req-aa73e81e-d653-495a-af20-a9f7092c99cc req-a68af173-b106-4e58-9a90-acb4b8a8d4ef service nova] Acquiring lock "refresh_cache-cede536f-eac5-4dd0-b4b5-74cfbd24300b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.559392] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Releasing lock "refresh_cache-cede536f-eac5-4dd0-b4b5-74cfbd24300b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.559980] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 625.560362] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.560991] env[62183]: DEBUG oslo_concurrency.lockutils [req-aa73e81e-d653-495a-af20-a9f7092c99cc req-a68af173-b106-4e58-9a90-acb4b8a8d4ef service nova] Acquired lock "refresh_cache-cede536f-eac5-4dd0-b4b5-74cfbd24300b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.560991] env[62183]: DEBUG nova.network.neutron [req-aa73e81e-d653-495a-af20-a9f7092c99cc req-a68af173-b106-4e58-9a90-acb4b8a8d4ef service nova] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Refreshing network info cache for port 60361efe-6879-402d-aa00-16d78c27baac {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.562136] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-983a174b-c491-469d-a4bb-99189f8ed03e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.574589] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af48fa0e-2f1b-4443-af28-91e579436819 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.597252] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cede536f-eac5-4dd0-b4b5-74cfbd24300b could not be found. [ 625.597530] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.597827] env[62183]: INFO nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 625.598096] env[62183]: DEBUG oslo.service.loopingcall [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.598307] env[62183]: DEBUG nova.compute.manager [-] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.598403] env[62183]: DEBUG nova.network.neutron [-] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.616343] env[62183]: DEBUG nova.network.neutron [-] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.677214] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Successfully created port: 4a93c18a-e699-47f0-9604-eeaf645a886d {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.767015] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 625.774810] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1a6e38e-c475-4485-bbfd-e26b58cd2ec5 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.774810] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.520s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.774810] env[62183]: INFO nova.compute.claims [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.082816] env[62183]: DEBUG nova.network.neutron [req-aa73e81e-d653-495a-af20-a9f7092c99cc req-a68af173-b106-4e58-9a90-acb4b8a8d4ef service nova] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.118768] env[62183]: DEBUG nova.network.neutron [-] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.165600] env[62183]: DEBUG nova.network.neutron [req-aa73e81e-d653-495a-af20-a9f7092c99cc req-a68af173-b106-4e58-9a90-acb4b8a8d4ef service nova] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.621618] env[62183]: INFO nova.compute.manager [-] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Took 1.02 seconds to deallocate network for instance. [ 626.624733] env[62183]: DEBUG nova.compute.claims [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.624912] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.668759] env[62183]: DEBUG oslo_concurrency.lockutils [req-aa73e81e-d653-495a-af20-a9f7092c99cc req-a68af173-b106-4e58-9a90-acb4b8a8d4ef service nova] Releasing lock "refresh_cache-cede536f-eac5-4dd0-b4b5-74cfbd24300b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.669037] env[62183]: DEBUG nova.compute.manager [req-aa73e81e-d653-495a-af20-a9f7092c99cc req-a68af173-b106-4e58-9a90-acb4b8a8d4ef service nova] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Received event network-vif-deleted-60361efe-6879-402d-aa00-16d78c27baac {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.730481] env[62183]: ERROR nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4a93c18a-e699-47f0-9604-eeaf645a886d, please check neutron logs for more information. [ 626.730481] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.730481] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.730481] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.730481] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.730481] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.730481] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.730481] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.730481] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.730481] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 626.730481] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.730481] env[62183]: ERROR nova.compute.manager raise self.value [ 626.730481] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.730481] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.730481] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.730481] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.730952] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.730952] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.730952] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4a93c18a-e699-47f0-9604-eeaf645a886d, please check neutron logs for more information. [ 626.730952] env[62183]: ERROR nova.compute.manager [ 626.730952] env[62183]: Traceback (most recent call last): [ 626.730952] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.730952] env[62183]: listener.cb(fileno) [ 626.730952] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.730952] env[62183]: result = function(*args, **kwargs) [ 626.730952] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.730952] env[62183]: return func(*args, **kwargs) [ 626.730952] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.730952] env[62183]: raise e [ 626.730952] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.730952] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 626.730952] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.730952] env[62183]: created_port_ids = self._update_ports_for_instance( [ 626.730952] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.730952] env[62183]: with excutils.save_and_reraise_exception(): [ 626.730952] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.730952] env[62183]: self.force_reraise() [ 626.730952] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.730952] env[62183]: raise self.value [ 626.730952] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.730952] env[62183]: updated_port = self._update_port( [ 626.730952] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.730952] env[62183]: _ensure_no_port_binding_failure(port) [ 626.730952] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.730952] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.731742] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 4a93c18a-e699-47f0-9604-eeaf645a886d, please check neutron logs for more information. [ 626.731742] env[62183]: Removing descriptor: 14 [ 626.784243] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.813976] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.814248] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.814409] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.814583] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.814725] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.814868] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.815080] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.815238] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.815402] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.815564] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.815729] env[62183]: DEBUG nova.virt.hardware [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.816571] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318ff951-a7a0-4f7a-8787-4ca3180ba829 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.826368] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6a155f-8238-4d75-9dfe-a8faa1aedb90 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.840919] env[62183]: ERROR nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4a93c18a-e699-47f0-9604-eeaf645a886d, please check neutron logs for more information. [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Traceback (most recent call last): [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] yield resources [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self.driver.spawn(context, instance, image_meta, [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] vm_ref = self.build_virtual_machine(instance, [ 626.840919] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] for vif in network_info: [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] return self._sync_wrapper(fn, *args, **kwargs) [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self.wait() [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self[:] = self._gt.wait() [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] return self._exit_event.wait() [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 626.841486] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] current.throw(*self._exc) [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] result = function(*args, **kwargs) [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] return func(*args, **kwargs) [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] raise e [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] nwinfo = self.network_api.allocate_for_instance( [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] created_port_ids = self._update_ports_for_instance( [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] with excutils.save_and_reraise_exception(): [ 626.841876] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self.force_reraise() [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] raise self.value [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] updated_port = self._update_port( [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] _ensure_no_port_binding_failure(port) [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] raise exception.PortBindingFailed(port_id=port['id']) [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] nova.exception.PortBindingFailed: Binding failed for port 4a93c18a-e699-47f0-9604-eeaf645a886d, please check neutron logs for more information. [ 626.842244] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] [ 626.842244] env[62183]: INFO nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Terminating instance [ 626.845540] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "refresh_cache-2c09e83a-e614-4c36-a3e1-35f6ed34e37e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.845540] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquired lock "refresh_cache-2c09e83a-e614-4c36-a3e1-35f6ed34e37e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.845540] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.165687] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d839c8fb-fc79-4800-bd7c-1163c8576089 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.173404] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f14123b-0da9-459e-bebf-c5e7c4c1fa17 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.203900] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509c5843-6d7e-42c2-be1a-d5b5fe9a8d02 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.211205] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0ca3ae-a18b-42b7-8d72-1b36c7f1603f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.224459] env[62183]: DEBUG nova.compute.provider_tree [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.369374] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.423741] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.546662] env[62183]: DEBUG nova.compute.manager [req-6b4fec49-c493-4b63-874b-15ff54d1bd33 req-522bccab-fe40-44e4-b1fb-4facb8cac0b7 service nova] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Received event network-changed-4a93c18a-e699-47f0-9604-eeaf645a886d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.546864] env[62183]: DEBUG nova.compute.manager [req-6b4fec49-c493-4b63-874b-15ff54d1bd33 req-522bccab-fe40-44e4-b1fb-4facb8cac0b7 service nova] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Refreshing instance network info cache due to event network-changed-4a93c18a-e699-47f0-9604-eeaf645a886d. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 627.547093] env[62183]: DEBUG oslo_concurrency.lockutils [req-6b4fec49-c493-4b63-874b-15ff54d1bd33 req-522bccab-fe40-44e4-b1fb-4facb8cac0b7 service nova] Acquiring lock "refresh_cache-2c09e83a-e614-4c36-a3e1-35f6ed34e37e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.727969] env[62183]: DEBUG nova.scheduler.client.report [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 627.929570] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Releasing lock "refresh_cache-2c09e83a-e614-4c36-a3e1-35f6ed34e37e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.930173] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 627.930442] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 627.930764] env[62183]: DEBUG oslo_concurrency.lockutils [req-6b4fec49-c493-4b63-874b-15ff54d1bd33 req-522bccab-fe40-44e4-b1fb-4facb8cac0b7 service nova] Acquired lock "refresh_cache-2c09e83a-e614-4c36-a3e1-35f6ed34e37e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.930938] env[62183]: DEBUG nova.network.neutron [req-6b4fec49-c493-4b63-874b-15ff54d1bd33 req-522bccab-fe40-44e4-b1fb-4facb8cac0b7 service nova] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Refreshing network info cache for port 4a93c18a-e699-47f0-9604-eeaf645a886d {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 627.932107] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-305dabfb-4ccc-4f8b-9d7a-bf06aa792e19 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.941387] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f6c6a7-f0ff-4575-8bf3-47dbc58e59df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.963909] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2c09e83a-e614-4c36-a3e1-35f6ed34e37e could not be found. [ 627.963909] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 627.963909] env[62183]: INFO nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 627.964303] env[62183]: DEBUG oslo.service.loopingcall [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 627.964520] env[62183]: DEBUG nova.compute.manager [-] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.964613] env[62183]: DEBUG nova.network.neutron [-] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 627.979583] env[62183]: DEBUG nova.network.neutron [-] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.233857] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.234285] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 628.236843] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.958s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.238257] env[62183]: INFO nova.compute.claims [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.451056] env[62183]: DEBUG nova.network.neutron [req-6b4fec49-c493-4b63-874b-15ff54d1bd33 req-522bccab-fe40-44e4-b1fb-4facb8cac0b7 service nova] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.482426] env[62183]: DEBUG nova.network.neutron [-] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.507537] env[62183]: DEBUG nova.network.neutron [req-6b4fec49-c493-4b63-874b-15ff54d1bd33 req-522bccab-fe40-44e4-b1fb-4facb8cac0b7 service nova] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.742526] env[62183]: DEBUG nova.compute.utils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.746180] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.746362] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 628.804211] env[62183]: DEBUG nova.policy [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc7a31885a37424aac4542af70afa679', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b65c4ef8c48f43c5a0af59686c8a6460', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.986897] env[62183]: INFO nova.compute.manager [-] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Took 1.02 seconds to deallocate network for instance. [ 628.988611] env[62183]: DEBUG nova.compute.claims [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 628.988794] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.009582] env[62183]: DEBUG oslo_concurrency.lockutils [req-6b4fec49-c493-4b63-874b-15ff54d1bd33 req-522bccab-fe40-44e4-b1fb-4facb8cac0b7 service nova] Releasing lock "refresh_cache-2c09e83a-e614-4c36-a3e1-35f6ed34e37e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.009797] env[62183]: DEBUG nova.compute.manager [req-6b4fec49-c493-4b63-874b-15ff54d1bd33 req-522bccab-fe40-44e4-b1fb-4facb8cac0b7 service nova] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Received event network-vif-deleted-4a93c18a-e699-47f0-9604-eeaf645a886d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.162643] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Successfully created port: fc56a552-1a30-46b3-8156-73c1a5322731 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.253438] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 629.633592] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554d00be-e391-4fd6-a7b0-a8792be97148 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.648200] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f715e6-4553-4888-9aa3-575322eb9556 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.675238] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbfc654-baa7-4e06-a6db-fc48f7f29e48 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.682351] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3181df22-bc8a-42fc-82b9-87c6900fe7d4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.695116] env[62183]: DEBUG nova.compute.provider_tree [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.984418] env[62183]: DEBUG nova.compute.manager [req-c74bbbb8-2681-40cb-b059-6d1d56729e21 req-f7a74896-c981-46ed-ad46-8016a7f450a6 service nova] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Received event network-changed-fc56a552-1a30-46b3-8156-73c1a5322731 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.984616] env[62183]: DEBUG nova.compute.manager [req-c74bbbb8-2681-40cb-b059-6d1d56729e21 req-f7a74896-c981-46ed-ad46-8016a7f450a6 service nova] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Refreshing instance network info cache due to event network-changed-fc56a552-1a30-46b3-8156-73c1a5322731. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 629.984824] env[62183]: DEBUG oslo_concurrency.lockutils [req-c74bbbb8-2681-40cb-b059-6d1d56729e21 req-f7a74896-c981-46ed-ad46-8016a7f450a6 service nova] Acquiring lock "refresh_cache-c1d8010b-14b0-4571-9d47-b0e05ebef3ed" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.984966] env[62183]: DEBUG oslo_concurrency.lockutils [req-c74bbbb8-2681-40cb-b059-6d1d56729e21 req-f7a74896-c981-46ed-ad46-8016a7f450a6 service nova] Acquired lock "refresh_cache-c1d8010b-14b0-4571-9d47-b0e05ebef3ed" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.985183] env[62183]: DEBUG nova.network.neutron [req-c74bbbb8-2681-40cb-b059-6d1d56729e21 req-f7a74896-c981-46ed-ad46-8016a7f450a6 service nova] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Refreshing network info cache for port fc56a552-1a30-46b3-8156-73c1a5322731 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 630.152048] env[62183]: ERROR nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fc56a552-1a30-46b3-8156-73c1a5322731, please check neutron logs for more information. [ 630.152048] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 630.152048] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.152048] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 630.152048] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.152048] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 630.152048] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.152048] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 630.152048] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.152048] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 630.152048] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.152048] env[62183]: ERROR nova.compute.manager raise self.value [ 630.152048] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.152048] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 630.152048] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.152048] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 630.152819] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.152819] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 630.152819] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fc56a552-1a30-46b3-8156-73c1a5322731, please check neutron logs for more information. [ 630.152819] env[62183]: ERROR nova.compute.manager [ 630.152819] env[62183]: Traceback (most recent call last): [ 630.152819] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 630.152819] env[62183]: listener.cb(fileno) [ 630.152819] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.152819] env[62183]: result = function(*args, **kwargs) [ 630.152819] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.152819] env[62183]: return func(*args, **kwargs) [ 630.152819] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.152819] env[62183]: raise e [ 630.152819] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.152819] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 630.152819] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.152819] env[62183]: created_port_ids = self._update_ports_for_instance( [ 630.152819] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.152819] env[62183]: with excutils.save_and_reraise_exception(): [ 630.152819] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.152819] env[62183]: self.force_reraise() [ 630.152819] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.152819] env[62183]: raise self.value [ 630.152819] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.152819] env[62183]: updated_port = self._update_port( [ 630.152819] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.152819] env[62183]: _ensure_no_port_binding_failure(port) [ 630.152819] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.152819] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 630.153548] env[62183]: nova.exception.PortBindingFailed: Binding failed for port fc56a552-1a30-46b3-8156-73c1a5322731, please check neutron logs for more information. [ 630.153548] env[62183]: Removing descriptor: 14 [ 630.198479] env[62183]: DEBUG nova.scheduler.client.report [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 630.262152] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 630.288761] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 630.289016] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 630.289181] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 630.289363] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 630.289519] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 630.289684] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 630.289900] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 630.290083] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 630.290267] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 630.290455] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 630.290633] env[62183]: DEBUG nova.virt.hardware [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.291490] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e6cfce-cb0d-449d-9ea9-308a1d601573 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.300373] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e43aae-a7b5-4187-997b-b41a4262d703 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.312723] env[62183]: ERROR nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fc56a552-1a30-46b3-8156-73c1a5322731, please check neutron logs for more information. [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Traceback (most recent call last): [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] yield resources [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self.driver.spawn(context, instance, image_meta, [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] vm_ref = self.build_virtual_machine(instance, [ 630.312723] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] for vif in network_info: [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] return self._sync_wrapper(fn, *args, **kwargs) [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self.wait() [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self[:] = self._gt.wait() [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] return self._exit_event.wait() [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 630.313066] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] current.throw(*self._exc) [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] result = function(*args, **kwargs) [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] return func(*args, **kwargs) [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] raise e [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] nwinfo = self.network_api.allocate_for_instance( [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] created_port_ids = self._update_ports_for_instance( [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] with excutils.save_and_reraise_exception(): [ 630.313415] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self.force_reraise() [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] raise self.value [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] updated_port = self._update_port( [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] _ensure_no_port_binding_failure(port) [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] raise exception.PortBindingFailed(port_id=port['id']) [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] nova.exception.PortBindingFailed: Binding failed for port fc56a552-1a30-46b3-8156-73c1a5322731, please check neutron logs for more information. [ 630.313832] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] [ 630.313832] env[62183]: INFO nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Terminating instance [ 630.315018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Acquiring lock "refresh_cache-c1d8010b-14b0-4571-9d47-b0e05ebef3ed" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.504205] env[62183]: DEBUG nova.network.neutron [req-c74bbbb8-2681-40cb-b059-6d1d56729e21 req-f7a74896-c981-46ed-ad46-8016a7f450a6 service nova] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.568206] env[62183]: DEBUG nova.network.neutron [req-c74bbbb8-2681-40cb-b059-6d1d56729e21 req-f7a74896-c981-46ed-ad46-8016a7f450a6 service nova] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.703775] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.704182] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.706722] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.665s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.706936] env[62183]: DEBUG nova.objects.instance [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lazy-loading 'resources' on Instance uuid a11cfd43-7516-4f98-8484-9ba2a0c5276d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 631.071290] env[62183]: DEBUG oslo_concurrency.lockutils [req-c74bbbb8-2681-40cb-b059-6d1d56729e21 req-f7a74896-c981-46ed-ad46-8016a7f450a6 service nova] Releasing lock "refresh_cache-c1d8010b-14b0-4571-9d47-b0e05ebef3ed" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.071703] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Acquired lock "refresh_cache-c1d8010b-14b0-4571-9d47-b0e05ebef3ed" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.071932] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.209598] env[62183]: DEBUG nova.compute.utils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.211507] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 631.211507] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 631.271581] env[62183]: DEBUG nova.policy [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f80808d61aa42f283d2d47631137403', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee3e2b1fecd74debb8362622a78039a2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.566343] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e1aef2-c934-43a6-9357-4e4dc7af0a89 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.574033] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974952ff-900b-4e92-a3e6-43587ee90b51 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.609030] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65042ae-dc70-4da6-a883-2f613b8c1e4e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.614128] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024fbcdf-bd1b-44a0-96a2-71821c2d751c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.627206] env[62183]: DEBUG nova.compute.provider_tree [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.629287] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.716663] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 631.720714] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.832587] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Successfully created port: 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 632.009755] env[62183]: DEBUG nova.compute.manager [req-1b25a038-e60e-4290-8152-30cc9f037391 req-21095534-8c6c-45c3-89db-3687d295d2ea service nova] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Received event network-vif-deleted-fc56a552-1a30-46b3-8156-73c1a5322731 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 632.133673] env[62183]: DEBUG nova.scheduler.client.report [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 632.226988] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Releasing lock "refresh_cache-c1d8010b-14b0-4571-9d47-b0e05ebef3ed" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.227632] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.227875] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.228363] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4ed166b8-4f8a-4c77-87d5-40ab72dbb044 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.237151] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81de4c76-30cf-49f7-b975-0c5941453598 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.260940] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c1d8010b-14b0-4571-9d47-b0e05ebef3ed could not be found. [ 632.260940] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.260940] env[62183]: INFO nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Took 0.03 seconds to destroy the instance on the hypervisor. [ 632.260940] env[62183]: DEBUG oslo.service.loopingcall [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.260940] env[62183]: DEBUG nova.compute.manager [-] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.260940] env[62183]: DEBUG nova.network.neutron [-] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.278424] env[62183]: DEBUG nova.network.neutron [-] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.638259] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.931s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.641238] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.797s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.666705] env[62183]: INFO nova.scheduler.client.report [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Deleted allocations for instance a11cfd43-7516-4f98-8484-9ba2a0c5276d [ 632.729783] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 632.764034] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.764667] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.765105] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.765723] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.766044] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.768839] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.768839] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.768839] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.768839] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.768839] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.769064] env[62183]: DEBUG nova.virt.hardware [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.769064] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f20fddd-fa81-4cd8-aea1-960a48f68f5d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.776931] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d04f8f-465e-459e-a0b6-4ae7c0ac2c1f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.781682] env[62183]: DEBUG nova.network.neutron [-] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.808925] env[62183]: ERROR nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf, please check neutron logs for more information. [ 632.808925] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.808925] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.808925] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.808925] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.808925] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.808925] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.808925] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.808925] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.808925] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 632.808925] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.808925] env[62183]: ERROR nova.compute.manager raise self.value [ 632.808925] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.808925] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.808925] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.808925] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.809320] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.809320] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.809320] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf, please check neutron logs for more information. [ 632.809320] env[62183]: ERROR nova.compute.manager [ 632.809320] env[62183]: Traceback (most recent call last): [ 632.809320] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.809320] env[62183]: listener.cb(fileno) [ 632.809320] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.809320] env[62183]: result = function(*args, **kwargs) [ 632.809320] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 632.809320] env[62183]: return func(*args, **kwargs) [ 632.809320] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.809320] env[62183]: raise e [ 632.809320] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.809320] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 632.809320] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.809320] env[62183]: created_port_ids = self._update_ports_for_instance( [ 632.809320] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.809320] env[62183]: with excutils.save_and_reraise_exception(): [ 632.809320] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.809320] env[62183]: self.force_reraise() [ 632.809320] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.809320] env[62183]: raise self.value [ 632.809320] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.809320] env[62183]: updated_port = self._update_port( [ 632.809320] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.809320] env[62183]: _ensure_no_port_binding_failure(port) [ 632.809320] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.809320] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.810018] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf, please check neutron logs for more information. [ 632.810018] env[62183]: Removing descriptor: 14 [ 632.810018] env[62183]: ERROR nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf, please check neutron logs for more information. [ 632.810018] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Traceback (most recent call last): [ 632.810018] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.810018] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] yield resources [ 632.810018] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.810018] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self.driver.spawn(context, instance, image_meta, [ 632.810018] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 632.810018] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.810018] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.810018] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] vm_ref = self.build_virtual_machine(instance, [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] for vif in network_info: [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] return self._sync_wrapper(fn, *args, **kwargs) [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self.wait() [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self[:] = self._gt.wait() [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] return self._exit_event.wait() [ 632.810305] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] result = hub.switch() [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] return self.greenlet.switch() [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] result = function(*args, **kwargs) [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] return func(*args, **kwargs) [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] raise e [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] nwinfo = self.network_api.allocate_for_instance( [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.810630] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] created_port_ids = self._update_ports_for_instance( [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] with excutils.save_and_reraise_exception(): [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self.force_reraise() [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] raise self.value [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] updated_port = self._update_port( [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] _ensure_no_port_binding_failure(port) [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.810925] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] raise exception.PortBindingFailed(port_id=port['id']) [ 632.811248] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] nova.exception.PortBindingFailed: Binding failed for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf, please check neutron logs for more information. [ 632.811248] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] [ 632.811248] env[62183]: INFO nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Terminating instance [ 632.811337] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "refresh_cache-b4a3b4f7-d726-4291-94ad-d75d656a9764" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.811462] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquired lock "refresh_cache-b4a3b4f7-d726-4291-94ad-d75d656a9764" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.811639] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.174728] env[62183]: DEBUG oslo_concurrency.lockutils [None req-46a9ec0a-e531-440b-9605-87712fa662a1 tempest-ServerDiagnosticsV248Test-1714655395 tempest-ServerDiagnosticsV248Test-1714655395-project-member] Lock "a11cfd43-7516-4f98-8484-9ba2a0c5276d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.557s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.284662] env[62183]: INFO nova.compute.manager [-] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Took 1.02 seconds to deallocate network for instance. [ 633.287265] env[62183]: DEBUG nova.compute.claims [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.287480] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.330597] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.413941] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.483124] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8e5015-6ee7-4ded-9a9b-97b35b19e6a6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.490674] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7463af-3bcf-49a7-9a7e-e97e333a8dd3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.520823] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66933db-a125-4ee1-92cf-487c9e63dc3a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.528480] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b997f5b9-2f0c-450a-9122-6c5f72ebe4e0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.542703] env[62183]: DEBUG nova.compute.provider_tree [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.918954] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Releasing lock "refresh_cache-b4a3b4f7-d726-4291-94ad-d75d656a9764" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.919414] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.919664] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 633.919956] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ec50b96-e207-450a-911f-28b311f035eb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.929138] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aafbc909-cd64-415a-820b-7ce513f4c229 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.951182] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b4a3b4f7-d726-4291-94ad-d75d656a9764 could not be found. [ 633.951370] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.951540] env[62183]: INFO nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Took 0.03 seconds to destroy the instance on the hypervisor. [ 633.951910] env[62183]: DEBUG oslo.service.loopingcall [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.952051] env[62183]: DEBUG nova.compute.manager [-] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.952148] env[62183]: DEBUG nova.network.neutron [-] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.966812] env[62183]: DEBUG nova.network.neutron [-] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.038622] env[62183]: DEBUG nova.compute.manager [req-787bd9d3-f955-4d1e-9d42-93cf7b43669b req-d2e49fae-dc20-4258-9faf-6511c4767211 service nova] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Received event network-changed-8eaca29e-ecf9-46d2-84ba-7c50bfa373cf {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.038622] env[62183]: DEBUG nova.compute.manager [req-787bd9d3-f955-4d1e-9d42-93cf7b43669b req-d2e49fae-dc20-4258-9faf-6511c4767211 service nova] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Refreshing instance network info cache due to event network-changed-8eaca29e-ecf9-46d2-84ba-7c50bfa373cf. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 634.038622] env[62183]: DEBUG oslo_concurrency.lockutils [req-787bd9d3-f955-4d1e-9d42-93cf7b43669b req-d2e49fae-dc20-4258-9faf-6511c4767211 service nova] Acquiring lock "refresh_cache-b4a3b4f7-d726-4291-94ad-d75d656a9764" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.038765] env[62183]: DEBUG oslo_concurrency.lockutils [req-787bd9d3-f955-4d1e-9d42-93cf7b43669b req-d2e49fae-dc20-4258-9faf-6511c4767211 service nova] Acquired lock "refresh_cache-b4a3b4f7-d726-4291-94ad-d75d656a9764" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.038870] env[62183]: DEBUG nova.network.neutron [req-787bd9d3-f955-4d1e-9d42-93cf7b43669b req-d2e49fae-dc20-4258-9faf-6511c4767211 service nova] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Refreshing network info cache for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 634.045561] env[62183]: DEBUG nova.scheduler.client.report [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 634.471036] env[62183]: DEBUG nova.network.neutron [-] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.551419] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.551912] env[62183]: ERROR nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 92cf027c-9f9f-4b98-b960-273b2ab787e1, please check neutron logs for more information. [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Traceback (most recent call last): [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self.driver.spawn(context, instance, image_meta, [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] vm_ref = self.build_virtual_machine(instance, [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.551912] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] for vif in network_info: [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] return self._sync_wrapper(fn, *args, **kwargs) [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self.wait() [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self[:] = self._gt.wait() [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] return self._exit_event.wait() [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] current.throw(*self._exc) [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.552273] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] result = function(*args, **kwargs) [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] return func(*args, **kwargs) [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] raise e [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] nwinfo = self.network_api.allocate_for_instance( [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] created_port_ids = self._update_ports_for_instance( [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] with excutils.save_and_reraise_exception(): [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] self.force_reraise() [ 634.552636] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.553103] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] raise self.value [ 634.553103] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.553103] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] updated_port = self._update_port( [ 634.553103] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.553103] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] _ensure_no_port_binding_failure(port) [ 634.553103] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.553103] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] raise exception.PortBindingFailed(port_id=port['id']) [ 634.553103] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] nova.exception.PortBindingFailed: Binding failed for port 92cf027c-9f9f-4b98-b960-273b2ab787e1, please check neutron logs for more information. [ 634.553103] env[62183]: ERROR nova.compute.manager [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] [ 634.553103] env[62183]: DEBUG nova.compute.utils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Binding failed for port 92cf027c-9f9f-4b98-b960-273b2ab787e1, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 634.553888] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.402s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.557051] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Build of instance 3ec662c7-940e-43fa-bbea-f3bc3243adf4 was re-scheduled: Binding failed for port 92cf027c-9f9f-4b98-b960-273b2ab787e1, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 634.557476] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 634.557695] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Acquiring lock "refresh_cache-3ec662c7-940e-43fa-bbea-f3bc3243adf4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.557839] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Acquired lock "refresh_cache-3ec662c7-940e-43fa-bbea-f3bc3243adf4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.557995] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 634.565893] env[62183]: DEBUG nova.network.neutron [req-787bd9d3-f955-4d1e-9d42-93cf7b43669b req-d2e49fae-dc20-4258-9faf-6511c4767211 service nova] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.626365] env[62183]: DEBUG nova.network.neutron [req-787bd9d3-f955-4d1e-9d42-93cf7b43669b req-d2e49fae-dc20-4258-9faf-6511c4767211 service nova] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.975603] env[62183]: INFO nova.compute.manager [-] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Took 1.02 seconds to deallocate network for instance. [ 634.976498] env[62183]: DEBUG nova.compute.claims [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.977636] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.076994] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.131743] env[62183]: DEBUG oslo_concurrency.lockutils [req-787bd9d3-f955-4d1e-9d42-93cf7b43669b req-d2e49fae-dc20-4258-9faf-6511c4767211 service nova] Releasing lock "refresh_cache-b4a3b4f7-d726-4291-94ad-d75d656a9764" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.132052] env[62183]: DEBUG nova.compute.manager [req-787bd9d3-f955-4d1e-9d42-93cf7b43669b req-d2e49fae-dc20-4258-9faf-6511c4767211 service nova] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Received event network-vif-deleted-8eaca29e-ecf9-46d2-84ba-7c50bfa373cf {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.186943] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.399354] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a22e99-a652-420a-9a81-9d42a63247d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.406822] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42f6598-02b5-4ea8-9e9d-ac210f1bb4c9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.437207] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5034cee-f0f1-4646-b608-cecb8ef92717 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.444686] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4243fc09-b85e-4cff-b654-4b0c9ced7633 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.458029] env[62183]: DEBUG nova.compute.provider_tree [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.689577] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Releasing lock "refresh_cache-3ec662c7-940e-43fa-bbea-f3bc3243adf4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.689838] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 635.690088] env[62183]: DEBUG nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.690309] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 635.705546] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.961144] env[62183]: DEBUG nova.scheduler.client.report [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 636.208466] env[62183]: DEBUG nova.network.neutron [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.466452] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.912s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.467355] env[62183]: ERROR nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 48058051-c983-4135-8170-3395eeadc4e3, please check neutron logs for more information. [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Traceback (most recent call last): [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self.driver.spawn(context, instance, image_meta, [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] vm_ref = self.build_virtual_machine(instance, [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.467355] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] for vif in network_info: [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] return self._sync_wrapper(fn, *args, **kwargs) [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self.wait() [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self[:] = self._gt.wait() [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] return self._exit_event.wait() [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] result = hub.switch() [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.469217] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] return self.greenlet.switch() [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] result = function(*args, **kwargs) [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] return func(*args, **kwargs) [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] raise e [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] nwinfo = self.network_api.allocate_for_instance( [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] created_port_ids = self._update_ports_for_instance( [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] with excutils.save_and_reraise_exception(): [ 636.469643] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] self.force_reraise() [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] raise self.value [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] updated_port = self._update_port( [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] _ensure_no_port_binding_failure(port) [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] raise exception.PortBindingFailed(port_id=port['id']) [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] nova.exception.PortBindingFailed: Binding failed for port 48058051-c983-4135-8170-3395eeadc4e3, please check neutron logs for more information. [ 636.469941] env[62183]: ERROR nova.compute.manager [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] [ 636.470210] env[62183]: DEBUG nova.compute.utils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Binding failed for port 48058051-c983-4135-8170-3395eeadc4e3, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.474018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.296s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.476798] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Build of instance 79c10e4e-98ab-4259-8ff5-72bfa39860be was re-scheduled: Binding failed for port 48058051-c983-4135-8170-3395eeadc4e3, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 636.481304] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 636.481304] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Acquiring lock "refresh_cache-79c10e4e-98ab-4259-8ff5-72bfa39860be" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.481304] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Acquired lock "refresh_cache-79c10e4e-98ab-4259-8ff5-72bfa39860be" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.481304] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.714261] env[62183]: INFO nova.compute.manager [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] [instance: 3ec662c7-940e-43fa-bbea-f3bc3243adf4] Took 1.02 seconds to deallocate network for instance. [ 637.022500] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.138796] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.433643] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957942cb-c332-4960-a415-62bfd80ad4d3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.441958] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e1b558-6eb9-4cde-8455-d808a9614e9b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.473704] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc143eb-d1b8-4156-86ed-d0ee39db31b4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.485258] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0b2ded-bac0-4ea9-998f-9f223b54c95a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.496927] env[62183]: DEBUG nova.compute.provider_tree [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.645285] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Releasing lock "refresh_cache-79c10e4e-98ab-4259-8ff5-72bfa39860be" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.645532] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 637.645703] env[62183]: DEBUG nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.645865] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.666975] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.747252] env[62183]: INFO nova.scheduler.client.report [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Deleted allocations for instance 3ec662c7-940e-43fa-bbea-f3bc3243adf4 [ 638.004294] env[62183]: DEBUG nova.scheduler.client.report [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 638.170076] env[62183]: DEBUG nova.network.neutron [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.263864] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b868eab3-e17e-4431-ae06-fbeaf72459cf tempest-ServersTestBootFromVolume-401660139 tempest-ServersTestBootFromVolume-401660139-project-member] Lock "3ec662c7-940e-43fa-bbea-f3bc3243adf4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.899s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.513947] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.042s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.514058] env[62183]: ERROR nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0b1c71e3-677d-4088-977d-208deb7610fe, please check neutron logs for more information. [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Traceback (most recent call last): [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self.driver.spawn(context, instance, image_meta, [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] vm_ref = self.build_virtual_machine(instance, [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.514058] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] for vif in network_info: [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] return self._sync_wrapper(fn, *args, **kwargs) [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self.wait() [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self[:] = self._gt.wait() [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] return self._exit_event.wait() [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] result = hub.switch() [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.514373] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] return self.greenlet.switch() [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] result = function(*args, **kwargs) [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] return func(*args, **kwargs) [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] raise e [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] nwinfo = self.network_api.allocate_for_instance( [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] created_port_ids = self._update_ports_for_instance( [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] with excutils.save_and_reraise_exception(): [ 638.514744] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] self.force_reraise() [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] raise self.value [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] updated_port = self._update_port( [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] _ensure_no_port_binding_failure(port) [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] raise exception.PortBindingFailed(port_id=port['id']) [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] nova.exception.PortBindingFailed: Binding failed for port 0b1c71e3-677d-4088-977d-208deb7610fe, please check neutron logs for more information. [ 638.515150] env[62183]: ERROR nova.compute.manager [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] [ 638.515474] env[62183]: DEBUG nova.compute.utils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Binding failed for port 0b1c71e3-677d-4088-977d-208deb7610fe, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.516337] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.210s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.516510] env[62183]: DEBUG nova.objects.instance [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62183) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 638.519192] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Build of instance ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65 was re-scheduled: Binding failed for port 0b1c71e3-677d-4088-977d-208deb7610fe, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.519619] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.519842] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Acquiring lock "refresh_cache-ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.520226] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Acquired lock "refresh_cache-ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.520226] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.673089] env[62183]: INFO nova.compute.manager [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] [instance: 79c10e4e-98ab-4259-8ff5-72bfa39860be] Took 1.03 seconds to deallocate network for instance. [ 638.766429] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.050874] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.206390] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.292303] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.531168] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8fa4a817-f603-4cc4-bd68-91aedbec56ac tempest-ServersAdmin275Test-835754308 tempest-ServersAdmin275Test-835754308-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.531168] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.072s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.706191] env[62183]: INFO nova.scheduler.client.report [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Deleted allocations for instance 79c10e4e-98ab-4259-8ff5-72bfa39860be [ 639.717262] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Releasing lock "refresh_cache-ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.722844] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 639.722844] env[62183]: DEBUG nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.722844] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.742871] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.225618] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0aa71dd9-7883-4e7a-a8bf-4dfa7db49d14 tempest-ServerDiagnosticsTest-1177361576 tempest-ServerDiagnosticsTest-1177361576-project-member] Lock "79c10e4e-98ab-4259-8ff5-72bfa39860be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.140s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.246503] env[62183]: DEBUG nova.network.neutron [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.437326] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791e7594-9c95-4a3c-87f6-415e1b69444c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.445809] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae4d089-6409-403b-88ba-def054843684 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.478823] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb83d16-495b-4c14-bb8c-f1037a16ad9f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.486369] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03937bc7-15b9-4f0a-b720-325ec523ddef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.502158] env[62183]: DEBUG nova.compute.provider_tree [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.729992] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 640.750034] env[62183]: INFO nova.compute.manager [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] [instance: ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65] Took 1.03 seconds to deallocate network for instance. [ 641.006981] env[62183]: DEBUG nova.scheduler.client.report [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 641.259533] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.513997] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.514680] env[62183]: ERROR nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 82ae4109-324e-4069-adad-1060e2f15a09, please check neutron logs for more information. [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Traceback (most recent call last): [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self.driver.spawn(context, instance, image_meta, [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] vm_ref = self.build_virtual_machine(instance, [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.514680] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] for vif in network_info: [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] return self._sync_wrapper(fn, *args, **kwargs) [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self.wait() [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self[:] = self._gt.wait() [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] return self._exit_event.wait() [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] result = hub.switch() [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 641.514984] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] return self.greenlet.switch() [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] result = function(*args, **kwargs) [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] return func(*args, **kwargs) [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] raise e [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] nwinfo = self.network_api.allocate_for_instance( [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] created_port_ids = self._update_ports_for_instance( [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] with excutils.save_and_reraise_exception(): [ 641.515851] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] self.force_reraise() [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] raise self.value [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] updated_port = self._update_port( [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] _ensure_no_port_binding_failure(port) [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] raise exception.PortBindingFailed(port_id=port['id']) [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] nova.exception.PortBindingFailed: Binding failed for port 82ae4109-324e-4069-adad-1060e2f15a09, please check neutron logs for more information. [ 641.516173] env[62183]: ERROR nova.compute.manager [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] [ 641.516482] env[62183]: DEBUG nova.compute.utils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Binding failed for port 82ae4109-324e-4069-adad-1060e2f15a09, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.516606] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.643s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.518858] env[62183]: INFO nova.compute.claims [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.521702] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Build of instance bed9c66e-a52f-4ec9-a190-a46c92547c09 was re-scheduled: Binding failed for port 82ae4109-324e-4069-adad-1060e2f15a09, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 641.522223] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 641.522507] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "refresh_cache-bed9c66e-a52f-4ec9-a190-a46c92547c09" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.522733] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "refresh_cache-bed9c66e-a52f-4ec9-a190-a46c92547c09" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.522945] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.791270] env[62183]: INFO nova.scheduler.client.report [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Deleted allocations for instance ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65 [ 642.049830] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.211512] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.309025] env[62183]: DEBUG oslo_concurrency.lockutils [None req-647dfbc3-dc0c-4d52-8dcd-1273637142aa tempest-InstanceActionsTestJSON-1531072542 tempest-InstanceActionsTestJSON-1531072542-project-member] Lock "ca8aaca3-7e6c-41ee-a640-b2aa1eaa2b65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.252s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.715296] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "refresh_cache-bed9c66e-a52f-4ec9-a190-a46c92547c09" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.715543] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 642.715725] env[62183]: DEBUG nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.715890] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 642.753759] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.810253] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 642.920018] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f444f5de-a70e-4449-93b1-ae210be7ddd1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.927841] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a79fb9-66b1-4a19-a96c-d74c816c72bb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.962874] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593a5480-5105-4c1b-a52d-de1462e3bd1c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.971221] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861cc548-c48f-499b-afc9-365494e5cb40 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.986379] env[62183]: DEBUG nova.compute.provider_tree [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.256592] env[62183]: DEBUG nova.network.neutron [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.350845] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.492874] env[62183]: DEBUG nova.scheduler.client.report [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 643.761881] env[62183]: INFO nova.compute.manager [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: bed9c66e-a52f-4ec9-a190-a46c92547c09] Took 1.04 seconds to deallocate network for instance. [ 643.997869] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.000056] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.002773] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 21.757s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.506765] env[62183]: DEBUG nova.compute.utils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 644.511564] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 644.511769] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 644.587868] env[62183]: DEBUG nova.policy [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e3a11f2fdf1c4a85bf4d0f889e9356e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5805ab93ca204bafaf3f278c6bbbe3ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 644.813432] env[62183]: INFO nova.scheduler.client.report [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleted allocations for instance bed9c66e-a52f-4ec9-a190-a46c92547c09 [ 644.971347] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Successfully created port: abc2f404-a0c3-4865-b54f-701e0a25728d {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.012749] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.049045] env[62183]: WARNING nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a43e24b0-c0f1-4e02-894d-0d2c8ed884d1 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 645.049045] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 650b8016-5a30-405a-b85c-6153228c9603 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 645.049045] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance cede536f-eac5-4dd0-b4b5-74cfbd24300b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 645.049045] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 2c09e83a-e614-4c36-a3e1-35f6ed34e37e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 645.049223] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance c1d8010b-14b0-4571-9d47-b0e05ebef3ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 645.049223] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance b4a3b4f7-d726-4291-94ad-d75d656a9764 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 645.049223] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 8e7fbc90-5be0-4baf-b371-1a22e84dee98 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 645.323163] env[62183]: DEBUG oslo_concurrency.lockutils [None req-137da6a8-56eb-424f-bc66-58cfe3bd728a tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "bed9c66e-a52f-4ec9-a190-a46c92547c09" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.762s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.552600] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 5f4edc62-4320-4aa4-ab91-ee77b757bb7c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.825598] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.029580] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 646.057258] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance f4b29d0e-488e-4d0c-b2cb-1321c40c87ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.070767] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.071183] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.071263] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.071381] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.071607] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.071790] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.071998] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.074278] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.074278] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.074278] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.074278] env[62183]: DEBUG nova.virt.hardware [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.074278] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d5027a-43f7-4ae1-8217-0a2febbb0731 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.083447] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0cad2f-1cef-4d27-9f61-5a0eee7971d2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.253820] env[62183]: DEBUG nova.compute.manager [req-19cbc5ad-2e12-4a19-b053-04cd8a2a1ae8 req-769e75b0-f916-4e80-b1e3-d3450fb7fe75 service nova] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Received event network-changed-abc2f404-a0c3-4865-b54f-701e0a25728d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 646.254031] env[62183]: DEBUG nova.compute.manager [req-19cbc5ad-2e12-4a19-b053-04cd8a2a1ae8 req-769e75b0-f916-4e80-b1e3-d3450fb7fe75 service nova] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Refreshing instance network info cache due to event network-changed-abc2f404-a0c3-4865-b54f-701e0a25728d. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 646.254314] env[62183]: DEBUG oslo_concurrency.lockutils [req-19cbc5ad-2e12-4a19-b053-04cd8a2a1ae8 req-769e75b0-f916-4e80-b1e3-d3450fb7fe75 service nova] Acquiring lock "refresh_cache-8e7fbc90-5be0-4baf-b371-1a22e84dee98" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.254389] env[62183]: DEBUG oslo_concurrency.lockutils [req-19cbc5ad-2e12-4a19-b053-04cd8a2a1ae8 req-769e75b0-f916-4e80-b1e3-d3450fb7fe75 service nova] Acquired lock "refresh_cache-8e7fbc90-5be0-4baf-b371-1a22e84dee98" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.254538] env[62183]: DEBUG nova.network.neutron [req-19cbc5ad-2e12-4a19-b053-04cd8a2a1ae8 req-769e75b0-f916-4e80-b1e3-d3450fb7fe75 service nova] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Refreshing network info cache for port abc2f404-a0c3-4865-b54f-701e0a25728d {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.354390] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.536610] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Acquiring lock "45adf688-561c-49c0-8395-a16bf4e43740" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.536857] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Lock "45adf688-561c-49c0-8395-a16bf4e43740" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.559965] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 0120298f-dc4c-4c13-ab71-70ccd6428dc9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.610447] env[62183]: ERROR nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port abc2f404-a0c3-4865-b54f-701e0a25728d, please check neutron logs for more information. [ 646.610447] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 646.610447] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.610447] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 646.610447] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.610447] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 646.610447] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.610447] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 646.610447] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.610447] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 646.610447] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.610447] env[62183]: ERROR nova.compute.manager raise self.value [ 646.610447] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.610447] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 646.610447] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.610447] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 646.610985] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.610985] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 646.610985] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port abc2f404-a0c3-4865-b54f-701e0a25728d, please check neutron logs for more information. [ 646.610985] env[62183]: ERROR nova.compute.manager [ 646.610985] env[62183]: Traceback (most recent call last): [ 646.610985] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 646.610985] env[62183]: listener.cb(fileno) [ 646.610985] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.610985] env[62183]: result = function(*args, **kwargs) [ 646.610985] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.610985] env[62183]: return func(*args, **kwargs) [ 646.610985] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.610985] env[62183]: raise e [ 646.610985] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.610985] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 646.610985] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.610985] env[62183]: created_port_ids = self._update_ports_for_instance( [ 646.610985] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.610985] env[62183]: with excutils.save_and_reraise_exception(): [ 646.610985] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.610985] env[62183]: self.force_reraise() [ 646.610985] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.610985] env[62183]: raise self.value [ 646.610985] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.610985] env[62183]: updated_port = self._update_port( [ 646.610985] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.610985] env[62183]: _ensure_no_port_binding_failure(port) [ 646.610985] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.610985] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 646.611876] env[62183]: nova.exception.PortBindingFailed: Binding failed for port abc2f404-a0c3-4865-b54f-701e0a25728d, please check neutron logs for more information. [ 646.611876] env[62183]: Removing descriptor: 14 [ 646.611876] env[62183]: ERROR nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port abc2f404-a0c3-4865-b54f-701e0a25728d, please check neutron logs for more information. [ 646.611876] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Traceback (most recent call last): [ 646.611876] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 646.611876] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] yield resources [ 646.611876] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.611876] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self.driver.spawn(context, instance, image_meta, [ 646.611876] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 646.611876] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.611876] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.611876] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] vm_ref = self.build_virtual_machine(instance, [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] for vif in network_info: [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] return self._sync_wrapper(fn, *args, **kwargs) [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self.wait() [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self[:] = self._gt.wait() [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] return self._exit_event.wait() [ 646.612236] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] result = hub.switch() [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] return self.greenlet.switch() [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] result = function(*args, **kwargs) [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] return func(*args, **kwargs) [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] raise e [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] nwinfo = self.network_api.allocate_for_instance( [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.612612] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] created_port_ids = self._update_ports_for_instance( [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] with excutils.save_and_reraise_exception(): [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self.force_reraise() [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] raise self.value [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] updated_port = self._update_port( [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] _ensure_no_port_binding_failure(port) [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.612995] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] raise exception.PortBindingFailed(port_id=port['id']) [ 646.613343] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] nova.exception.PortBindingFailed: Binding failed for port abc2f404-a0c3-4865-b54f-701e0a25728d, please check neutron logs for more information. [ 646.613343] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] [ 646.613343] env[62183]: INFO nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Terminating instance [ 646.614254] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "refresh_cache-8e7fbc90-5be0-4baf-b371-1a22e84dee98" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.787787] env[62183]: DEBUG nova.network.neutron [req-19cbc5ad-2e12-4a19-b053-04cd8a2a1ae8 req-769e75b0-f916-4e80-b1e3-d3450fb7fe75 service nova] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.892273] env[62183]: DEBUG nova.network.neutron [req-19cbc5ad-2e12-4a19-b053-04cd8a2a1ae8 req-769e75b0-f916-4e80-b1e3-d3450fb7fe75 service nova] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.063156] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.303928] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.304174] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.395053] env[62183]: DEBUG oslo_concurrency.lockutils [req-19cbc5ad-2e12-4a19-b053-04cd8a2a1ae8 req-769e75b0-f916-4e80-b1e3-d3450fb7fe75 service nova] Releasing lock "refresh_cache-8e7fbc90-5be0-4baf-b371-1a22e84dee98" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.396249] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquired lock "refresh_cache-8e7fbc90-5be0-4baf-b371-1a22e84dee98" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.396249] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.567593] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance f67ab6cd-410b-4ed3-890b-68f3b0f41ec7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.915701] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.018256] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.072540] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance ee99cd66-908e-448f-9cce-ebe78a32214d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.347907] env[62183]: DEBUG nova.compute.manager [req-6a2ffa19-a509-48cd-b938-56a33998e492 req-c1bb5a46-edc6-4fd2-a245-38b515d8963a service nova] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Received event network-vif-deleted-abc2f404-a0c3-4865-b54f-701e0a25728d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.521548] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Releasing lock "refresh_cache-8e7fbc90-5be0-4baf-b371-1a22e84dee98" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.523038] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.523038] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.523038] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a3e2fbf-a10b-411e-9ad9-281edeb9be16 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.531143] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f554cb04-89e1-467d-b216-e8b5ef965acc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.556452] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8e7fbc90-5be0-4baf-b371-1a22e84dee98 could not be found. [ 648.556681] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.556869] env[62183]: INFO nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Took 0.03 seconds to destroy the instance on the hypervisor. [ 648.557130] env[62183]: DEBUG oslo.service.loopingcall [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.557363] env[62183]: DEBUG nova.compute.manager [-] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.557460] env[62183]: DEBUG nova.network.neutron [-] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.572018] env[62183]: DEBUG nova.network.neutron [-] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.577827] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.075462] env[62183]: DEBUG nova.network.neutron [-] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.081828] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 4af88a37-5aa2-47af-9dd9-8233b1bbf077 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.578878] env[62183]: INFO nova.compute.manager [-] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Took 1.02 seconds to deallocate network for instance. [ 649.581379] env[62183]: DEBUG nova.compute.claims [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.581564] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.584431] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 8397279a-7a43-4fb0-bc27-17e17e258527 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.754386] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Acquiring lock "4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.754613] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Lock "4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.087166] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 1ad515a9-38ab-4094-9d71-c845dcb72cb6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.590929] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance c2182423-e7fe-4ae2-be94-e88a1e49bab9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.093523] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 541531de-d96c-47b5-a980-34dfc3e904aa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.600026] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 82816947-cb4f-4ad7-ad44-1403db312cf8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.102977] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a1cabc70-e6c0-4c9b-8302-6ed8805f2182 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.606247] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 3cb4be60-b626-4cc9-91eb-2bd9b778f153 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.109274] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance f25ef564-b649-4557-9847-b5d994079d20 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.612789] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance b0761551-b2a3-4a38-a32d-f94a36c47d94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.115911] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 430dcd52-17ce-4710-b9c8-39c28410e356 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.116253] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 654.116439] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 654.497602] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6b99eb-6785-4920-b56c-7d4cb8c96a0d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.505259] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9e9b14-1402-4a1d-9623-49346457be97 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.536272] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881eb4c9-aa01-462e-8b23-47b084644c0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.543639] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5277a34d-4606-4080-b5cc-9fe765d6723e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.556618] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.060052] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 655.565262] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 655.565262] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.562s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.565262] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.024s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.565456] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.567771] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.865s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.666331] env[62183]: INFO nova.scheduler.client.report [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Deleted allocations for instance a43e24b0-c0f1-4e02-894d-0d2c8ed884d1 [ 656.174264] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e7d60d13-d1cc-4519-8271-94259e86e9a0 tempest-ServersAdmin275Test-746201289 tempest-ServersAdmin275Test-746201289-project-member] Lock "a43e24b0-c0f1-4e02-894d-0d2c8ed884d1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.319s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.366523] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a714dba-8060-4e33-9ecf-b22c50cf7ff5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.374630] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22072a70-69bb-4601-bc1c-caf2691d265c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.407019] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7678484a-86f8-4a5e-bd61-a1c7b289cfe4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.416776] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae64840d-2ebe-46b0-a845-8155288cf855 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.431444] env[62183]: DEBUG nova.compute.provider_tree [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.934318] env[62183]: DEBUG nova.scheduler.client.report [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 657.439751] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.872s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.440468] env[62183]: ERROR nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691, please check neutron logs for more information. [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] Traceback (most recent call last): [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self.driver.spawn(context, instance, image_meta, [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] vm_ref = self.build_virtual_machine(instance, [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.440468] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] for vif in network_info: [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] return self._sync_wrapper(fn, *args, **kwargs) [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self.wait() [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self[:] = self._gt.wait() [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] return self._exit_event.wait() [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] current.throw(*self._exc) [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.440801] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] result = function(*args, **kwargs) [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] return func(*args, **kwargs) [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] raise e [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] nwinfo = self.network_api.allocate_for_instance( [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] created_port_ids = self._update_ports_for_instance( [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] with excutils.save_and_reraise_exception(): [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] self.force_reraise() [ 657.441119] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.441476] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] raise self.value [ 657.441476] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.441476] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] updated_port = self._update_port( [ 657.441476] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.441476] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] _ensure_no_port_binding_failure(port) [ 657.441476] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.441476] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] raise exception.PortBindingFailed(port_id=port['id']) [ 657.441476] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] nova.exception.PortBindingFailed: Binding failed for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691, please check neutron logs for more information. [ 657.441476] env[62183]: ERROR nova.compute.manager [instance: 650b8016-5a30-405a-b85c-6153228c9603] [ 657.441476] env[62183]: DEBUG nova.compute.utils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Binding failed for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.442528] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.818s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.445894] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Build of instance 650b8016-5a30-405a-b85c-6153228c9603 was re-scheduled: Binding failed for port 1655cefc-8ccb-4fa0-adcf-d4a850b8f691, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 657.446663] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 657.447052] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquiring lock "refresh_cache-650b8016-5a30-405a-b85c-6153228c9603" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.447215] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Acquired lock "refresh_cache-650b8016-5a30-405a-b85c-6153228c9603" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.447395] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.974989] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.045656] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.290178] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a87f363-ce86-42fc-8868-0759860c454a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.297430] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7afa6a-f7aa-4851-8ff5-bbd22c18ac98 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.326078] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd217445-09e9-4e01-b0b0-079cb5c86e80 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.332899] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6d2b93-439b-46c0-8d46-d5bdf32f6b0f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.348024] env[62183]: DEBUG nova.compute.provider_tree [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.551427] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Releasing lock "refresh_cache-650b8016-5a30-405a-b85c-6153228c9603" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.551765] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 658.551887] env[62183]: DEBUG nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.552072] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.572117] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.851335] env[62183]: DEBUG nova.scheduler.client.report [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 659.074601] env[62183]: DEBUG nova.network.neutron [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.359019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.359019] env[62183]: ERROR nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 60361efe-6879-402d-aa00-16d78c27baac, please check neutron logs for more information. [ 659.359019] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Traceback (most recent call last): [ 659.359019] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.359019] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self.driver.spawn(context, instance, image_meta, [ 659.359019] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 659.359019] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.359019] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.359019] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] vm_ref = self.build_virtual_machine(instance, [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] for vif in network_info: [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] return self._sync_wrapper(fn, *args, **kwargs) [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self.wait() [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self[:] = self._gt.wait() [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] return self._exit_event.wait() [ 659.359364] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] result = hub.switch() [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] return self.greenlet.switch() [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] result = function(*args, **kwargs) [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] return func(*args, **kwargs) [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] raise e [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] nwinfo = self.network_api.allocate_for_instance( [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.359739] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] created_port_ids = self._update_ports_for_instance( [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] with excutils.save_and_reraise_exception(): [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] self.force_reraise() [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] raise self.value [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] updated_port = self._update_port( [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] _ensure_no_port_binding_failure(port) [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.360247] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] raise exception.PortBindingFailed(port_id=port['id']) [ 659.360579] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] nova.exception.PortBindingFailed: Binding failed for port 60361efe-6879-402d-aa00-16d78c27baac, please check neutron logs for more information. [ 659.360579] env[62183]: ERROR nova.compute.manager [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] [ 659.360579] env[62183]: DEBUG nova.compute.utils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Binding failed for port 60361efe-6879-402d-aa00-16d78c27baac, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 659.360579] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.370s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.365019] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Build of instance cede536f-eac5-4dd0-b4b5-74cfbd24300b was re-scheduled: Binding failed for port 60361efe-6879-402d-aa00-16d78c27baac, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 659.365019] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 659.365019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "refresh_cache-cede536f-eac5-4dd0-b4b5-74cfbd24300b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.365019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquired lock "refresh_cache-cede536f-eac5-4dd0-b4b5-74cfbd24300b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.365329] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.578422] env[62183]: INFO nova.compute.manager [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] [instance: 650b8016-5a30-405a-b85c-6153228c9603] Took 1.03 seconds to deallocate network for instance. [ 659.898174] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.970369] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.278457] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab08598-a581-4496-91c9-2cc5ebd537ce {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.286015] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9460c3e7-0fa1-4c06-a5c8-a2f4e0fd9270 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.315022] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e45730-5ff3-46dd-aab0-e94f8a904dbf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.323481] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6883dee-66ae-422a-ab4c-033a4f580eed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.335215] env[62183]: DEBUG nova.compute.provider_tree [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.472560] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Releasing lock "refresh_cache-cede536f-eac5-4dd0-b4b5-74cfbd24300b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.472809] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 660.472984] env[62183]: DEBUG nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.473168] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.491820] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.612123] env[62183]: INFO nova.scheduler.client.report [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Deleted allocations for instance 650b8016-5a30-405a-b85c-6153228c9603 [ 660.838759] env[62183]: DEBUG nova.scheduler.client.report [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 660.997590] env[62183]: DEBUG nova.network.neutron [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.121261] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cd94e2ac-5a17-432b-b906-ed13b332d768 tempest-MigrationsAdminTest-797487164 tempest-MigrationsAdminTest-797487164-project-member] Lock "650b8016-5a30-405a-b85c-6153228c9603" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.995s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.343714] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.344374] env[62183]: ERROR nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4a93c18a-e699-47f0-9604-eeaf645a886d, please check neutron logs for more information. [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Traceback (most recent call last): [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self.driver.spawn(context, instance, image_meta, [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] vm_ref = self.build_virtual_machine(instance, [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.344374] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] for vif in network_info: [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] return self._sync_wrapper(fn, *args, **kwargs) [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self.wait() [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self[:] = self._gt.wait() [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] return self._exit_event.wait() [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] current.throw(*self._exc) [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.344672] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] result = function(*args, **kwargs) [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] return func(*args, **kwargs) [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] raise e [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] nwinfo = self.network_api.allocate_for_instance( [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] created_port_ids = self._update_ports_for_instance( [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] with excutils.save_and_reraise_exception(): [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] self.force_reraise() [ 661.344980] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.345345] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] raise self.value [ 661.345345] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 661.345345] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] updated_port = self._update_port( [ 661.345345] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.345345] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] _ensure_no_port_binding_failure(port) [ 661.345345] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.345345] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] raise exception.PortBindingFailed(port_id=port['id']) [ 661.345345] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] nova.exception.PortBindingFailed: Binding failed for port 4a93c18a-e699-47f0-9604-eeaf645a886d, please check neutron logs for more information. [ 661.345345] env[62183]: ERROR nova.compute.manager [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] [ 661.345345] env[62183]: DEBUG nova.compute.utils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Binding failed for port 4a93c18a-e699-47f0-9604-eeaf645a886d, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 661.346361] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.059s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.349875] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Build of instance 2c09e83a-e614-4c36-a3e1-35f6ed34e37e was re-scheduled: Binding failed for port 4a93c18a-e699-47f0-9604-eeaf645a886d, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 661.350304] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 661.350523] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquiring lock "refresh_cache-2c09e83a-e614-4c36-a3e1-35f6ed34e37e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.350727] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Acquired lock "refresh_cache-2c09e83a-e614-4c36-a3e1-35f6ed34e37e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.351024] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 661.500863] env[62183]: INFO nova.compute.manager [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: cede536f-eac5-4dd0-b4b5-74cfbd24300b] Took 1.03 seconds to deallocate network for instance. [ 661.627299] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.878195] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.994123] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.158435] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.219618] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490d2117-f1e6-40b9-a3c4-87aff92d7055 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.228341] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e859ee4-9f05-4037-bab0-4a683cad4e81 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.268254] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1092bc63-7543-43b8-9991-e2d8e0c63ea5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.276924] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638da826-fd9a-4dfd-8892-043ccc20f378 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.291611] env[62183]: DEBUG nova.compute.provider_tree [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.498467] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Releasing lock "refresh_cache-2c09e83a-e614-4c36-a3e1-35f6ed34e37e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.498930] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 662.499171] env[62183]: DEBUG nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.499354] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 662.521600] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.533446] env[62183]: INFO nova.scheduler.client.report [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Deleted allocations for instance cede536f-eac5-4dd0-b4b5-74cfbd24300b [ 662.794979] env[62183]: DEBUG nova.scheduler.client.report [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 663.023969] env[62183]: DEBUG nova.network.neutron [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.043877] env[62183]: DEBUG oslo_concurrency.lockutils [None req-77044f48-1781-4117-9cf3-4f57f8492a50 tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "cede536f-eac5-4dd0-b4b5-74cfbd24300b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.272s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.301646] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.955s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.302343] env[62183]: ERROR nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fc56a552-1a30-46b3-8156-73c1a5322731, please check neutron logs for more information. [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Traceback (most recent call last): [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self.driver.spawn(context, instance, image_meta, [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] vm_ref = self.build_virtual_machine(instance, [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.302343] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] for vif in network_info: [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] return self._sync_wrapper(fn, *args, **kwargs) [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self.wait() [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self[:] = self._gt.wait() [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] return self._exit_event.wait() [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] current.throw(*self._exc) [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.302643] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] result = function(*args, **kwargs) [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] return func(*args, **kwargs) [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] raise e [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] nwinfo = self.network_api.allocate_for_instance( [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] created_port_ids = self._update_ports_for_instance( [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] with excutils.save_and_reraise_exception(): [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] self.force_reraise() [ 663.303088] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.303409] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] raise self.value [ 663.303409] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.303409] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] updated_port = self._update_port( [ 663.303409] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.303409] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] _ensure_no_port_binding_failure(port) [ 663.303409] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.303409] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] raise exception.PortBindingFailed(port_id=port['id']) [ 663.303409] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] nova.exception.PortBindingFailed: Binding failed for port fc56a552-1a30-46b3-8156-73c1a5322731, please check neutron logs for more information. [ 663.303409] env[62183]: ERROR nova.compute.manager [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] [ 663.304713] env[62183]: DEBUG nova.compute.utils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Binding failed for port fc56a552-1a30-46b3-8156-73c1a5322731, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.307903] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.331s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.311258] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Build of instance c1d8010b-14b0-4571-9d47-b0e05ebef3ed was re-scheduled: Binding failed for port fc56a552-1a30-46b3-8156-73c1a5322731, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.311700] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.311938] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Acquiring lock "refresh_cache-c1d8010b-14b0-4571-9d47-b0e05ebef3ed" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.312645] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Acquired lock "refresh_cache-c1d8010b-14b0-4571-9d47-b0e05ebef3ed" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.312645] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.527238] env[62183]: INFO nova.compute.manager [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] [instance: 2c09e83a-e614-4c36-a3e1-35f6ed34e37e] Took 1.03 seconds to deallocate network for instance. [ 663.549710] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 663.845506] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.949733] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.082108] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.294070] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4557e562-3d03-472f-8c82-5ff5903b0490 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.302117] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4f654a-86e2-482e-8fc3-b3b6496d4049 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.333823] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87975747-6139-46e8-a033-7cab4069e298 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.340350] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4db8fa9-59fd-41fb-925a-0469d655a8cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.356398] env[62183]: DEBUG nova.compute.provider_tree [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.452844] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Releasing lock "refresh_cache-c1d8010b-14b0-4571-9d47-b0e05ebef3ed" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.453304] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.453501] env[62183]: DEBUG nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.454126] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.471473] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.563312] env[62183]: INFO nova.scheduler.client.report [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Deleted allocations for instance 2c09e83a-e614-4c36-a3e1-35f6ed34e37e [ 664.861171] env[62183]: DEBUG nova.scheduler.client.report [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 664.979322] env[62183]: DEBUG nova.network.neutron [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.074339] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8dfc47e4-a8e1-4fde-8d8e-01c3fa77ee2e tempest-ListImageFiltersTestJSON-421638589 tempest-ListImageFiltersTestJSON-421638589-project-member] Lock "2c09e83a-e614-4c36-a3e1-35f6ed34e37e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.499s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.365510] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.368500] env[62183]: ERROR nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf, please check neutron logs for more information. [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Traceback (most recent call last): [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self.driver.spawn(context, instance, image_meta, [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] vm_ref = self.build_virtual_machine(instance, [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.368500] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] for vif in network_info: [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] return self._sync_wrapper(fn, *args, **kwargs) [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self.wait() [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self[:] = self._gt.wait() [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] return self._exit_event.wait() [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] result = hub.switch() [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.369237] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] return self.greenlet.switch() [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] result = function(*args, **kwargs) [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] return func(*args, **kwargs) [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] raise e [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] nwinfo = self.network_api.allocate_for_instance( [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] created_port_ids = self._update_ports_for_instance( [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] with excutils.save_and_reraise_exception(): [ 665.369541] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] self.force_reraise() [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] raise self.value [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] updated_port = self._update_port( [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] _ensure_no_port_binding_failure(port) [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] raise exception.PortBindingFailed(port_id=port['id']) [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] nova.exception.PortBindingFailed: Binding failed for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf, please check neutron logs for more information. [ 665.369850] env[62183]: ERROR nova.compute.manager [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] [ 665.370174] env[62183]: DEBUG nova.compute.utils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Binding failed for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.370384] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.078s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.372263] env[62183]: INFO nova.compute.claims [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.375198] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Build of instance b4a3b4f7-d726-4291-94ad-d75d656a9764 was re-scheduled: Binding failed for port 8eaca29e-ecf9-46d2-84ba-7c50bfa373cf, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 665.375320] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 665.376155] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "refresh_cache-b4a3b4f7-d726-4291-94ad-d75d656a9764" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.376155] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquired lock "refresh_cache-b4a3b4f7-d726-4291-94ad-d75d656a9764" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.376155] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.484218] env[62183]: INFO nova.compute.manager [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] [instance: c1d8010b-14b0-4571-9d47-b0e05ebef3ed] Took 1.03 seconds to deallocate network for instance. [ 665.546087] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "9278c981-042e-41a2-a4f8-d42a8e6fac39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.546493] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "9278c981-042e-41a2-a4f8-d42a8e6fac39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.576959] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.908121] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.984934] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.016181] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "3b231f56-5127-4ff9-9031-06bfd59a57d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.018115] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "3b231f56-5127-4ff9-9031-06bfd59a57d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.105550] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.476989] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Acquiring lock "3fef2148-244d-4575-8e3a-93bf70c7dd7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.478247] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Lock "3fef2148-244d-4575-8e3a-93bf70c7dd7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.488379] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Releasing lock "refresh_cache-b4a3b4f7-d726-4291-94ad-d75d656a9764" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.488629] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 666.488782] env[62183]: DEBUG nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.488963] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.514149] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.517847] env[62183]: INFO nova.scheduler.client.report [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Deleted allocations for instance c1d8010b-14b0-4571-9d47-b0e05ebef3ed [ 666.840925] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3022ce4a-7b1e-4d94-bf2d-dab7edb8e319 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.848829] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e55f90f2-5d4a-4887-84d7-91ef704f72b5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.887451] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26eb4e7d-1dbe-40e2-9ff6-172cb70bc01d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.897216] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244f8250-30f8-4586-b096-50c8ad0a90cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.912761] env[62183]: DEBUG nova.compute.provider_tree [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.016996] env[62183]: DEBUG nova.network.neutron [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.025347] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3b54dc5d-1a86-47dc-95a6-cd1ea4716de0 tempest-ServerDiagnosticsNegativeTest-1681313975 tempest-ServerDiagnosticsNegativeTest-1681313975-project-member] Lock "c1d8010b-14b0-4571-9d47-b0e05ebef3ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.934s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.416427] env[62183]: DEBUG nova.scheduler.client.report [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 667.520330] env[62183]: INFO nova.compute.manager [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: b4a3b4f7-d726-4291-94ad-d75d656a9764] Took 1.03 seconds to deallocate network for instance. [ 667.529029] env[62183]: DEBUG nova.compute.manager [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.923156] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.924032] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 667.926198] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.667s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.928290] env[62183]: INFO nova.compute.claims [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 668.058403] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.431870] env[62183]: DEBUG nova.compute.utils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.435740] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 668.435972] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 668.488914] env[62183]: DEBUG nova.policy [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70c6875c5ddf4743bb48d8c66ca6c076', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4f739b5563b479eaaef1d5be5dcd847', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 668.559220] env[62183]: INFO nova.scheduler.client.report [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Deleted allocations for instance b4a3b4f7-d726-4291-94ad-d75d656a9764 [ 668.794844] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Successfully created port: efcc7c0d-32ee-4834-84b0-a32619083b33 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.937195] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 669.070022] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cbec8bf6-817c-4a7b-ac97-135f66f2ade8 tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "b4a3b4f7-d726-4291-94ad-d75d656a9764" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.676s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.373742] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e754e39-5e83-42f1-b789-93b533bee4fe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.382981] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c42cab-88bb-4be5-9753-2a867000cd61 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.415612] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91efec46-c829-4be7-9ded-bcc5c9d64e81 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.423283] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29021af-f05e-4605-97d9-95325f0eafdb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.443235] env[62183]: DEBUG nova.compute.provider_tree [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.573601] env[62183]: DEBUG nova.compute.manager [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 669.852664] env[62183]: DEBUG nova.compute.manager [req-d932f48b-62f7-4a78-9259-b5ce0eaa78dd req-0b1a1d21-4086-4d26-a401-46b862979037 service nova] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Received event network-changed-efcc7c0d-32ee-4834-84b0-a32619083b33 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.852664] env[62183]: DEBUG nova.compute.manager [req-d932f48b-62f7-4a78-9259-b5ce0eaa78dd req-0b1a1d21-4086-4d26-a401-46b862979037 service nova] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Refreshing instance network info cache due to event network-changed-efcc7c0d-32ee-4834-84b0-a32619083b33. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.852664] env[62183]: DEBUG oslo_concurrency.lockutils [req-d932f48b-62f7-4a78-9259-b5ce0eaa78dd req-0b1a1d21-4086-4d26-a401-46b862979037 service nova] Acquiring lock "refresh_cache-5f4edc62-4320-4aa4-ab91-ee77b757bb7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.852664] env[62183]: DEBUG oslo_concurrency.lockutils [req-d932f48b-62f7-4a78-9259-b5ce0eaa78dd req-0b1a1d21-4086-4d26-a401-46b862979037 service nova] Acquired lock "refresh_cache-5f4edc62-4320-4aa4-ab91-ee77b757bb7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.852810] env[62183]: DEBUG nova.network.neutron [req-d932f48b-62f7-4a78-9259-b5ce0eaa78dd req-0b1a1d21-4086-4d26-a401-46b862979037 service nova] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Refreshing network info cache for port efcc7c0d-32ee-4834-84b0-a32619083b33 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.921928] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "48909506-5abd-4dfd-9aac-f3c1ef0f8250" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.923714] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "48909506-5abd-4dfd-9aac-f3c1ef0f8250" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.944909] env[62183]: DEBUG nova.scheduler.client.report [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 669.955777] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 669.981270] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.981771] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.982021] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.982255] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.982433] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.982626] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.982890] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.983736] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.983736] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.983736] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.983886] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.985118] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad49066-e1bc-4558-a383-12476d49137e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.994595] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae86da5f-e492-4f08-a6f4-d74e18baeb90 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.046512] env[62183]: ERROR nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port efcc7c0d-32ee-4834-84b0-a32619083b33, please check neutron logs for more information. [ 670.046512] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 670.046512] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.046512] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 670.046512] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.046512] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 670.046512] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.046512] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 670.046512] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.046512] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 670.046512] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.046512] env[62183]: ERROR nova.compute.manager raise self.value [ 670.046512] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.046512] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 670.046512] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.046512] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 670.046972] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.046972] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 670.046972] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port efcc7c0d-32ee-4834-84b0-a32619083b33, please check neutron logs for more information. [ 670.046972] env[62183]: ERROR nova.compute.manager [ 670.046972] env[62183]: Traceback (most recent call last): [ 670.046972] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 670.046972] env[62183]: listener.cb(fileno) [ 670.046972] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.046972] env[62183]: result = function(*args, **kwargs) [ 670.046972] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.046972] env[62183]: return func(*args, **kwargs) [ 670.046972] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.046972] env[62183]: raise e [ 670.046972] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.046972] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 670.046972] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.046972] env[62183]: created_port_ids = self._update_ports_for_instance( [ 670.046972] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.046972] env[62183]: with excutils.save_and_reraise_exception(): [ 670.046972] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.046972] env[62183]: self.force_reraise() [ 670.046972] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.046972] env[62183]: raise self.value [ 670.046972] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.046972] env[62183]: updated_port = self._update_port( [ 670.046972] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.046972] env[62183]: _ensure_no_port_binding_failure(port) [ 670.046972] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.046972] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 670.047738] env[62183]: nova.exception.PortBindingFailed: Binding failed for port efcc7c0d-32ee-4834-84b0-a32619083b33, please check neutron logs for more information. [ 670.047738] env[62183]: Removing descriptor: 14 [ 670.047738] env[62183]: ERROR nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port efcc7c0d-32ee-4834-84b0-a32619083b33, please check neutron logs for more information. [ 670.047738] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Traceback (most recent call last): [ 670.047738] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 670.047738] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] yield resources [ 670.047738] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.047738] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self.driver.spawn(context, instance, image_meta, [ 670.047738] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 670.047738] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.047738] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.047738] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] vm_ref = self.build_virtual_machine(instance, [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] for vif in network_info: [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] return self._sync_wrapper(fn, *args, **kwargs) [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self.wait() [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self[:] = self._gt.wait() [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] return self._exit_event.wait() [ 670.048226] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] result = hub.switch() [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] return self.greenlet.switch() [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] result = function(*args, **kwargs) [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] return func(*args, **kwargs) [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] raise e [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] nwinfo = self.network_api.allocate_for_instance( [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.048563] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] created_port_ids = self._update_ports_for_instance( [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] with excutils.save_and_reraise_exception(): [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self.force_reraise() [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] raise self.value [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] updated_port = self._update_port( [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] _ensure_no_port_binding_failure(port) [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.048888] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] raise exception.PortBindingFailed(port_id=port['id']) [ 670.049210] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] nova.exception.PortBindingFailed: Binding failed for port efcc7c0d-32ee-4834-84b0-a32619083b33, please check neutron logs for more information. [ 670.049210] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] [ 670.049210] env[62183]: INFO nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Terminating instance [ 670.050039] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "refresh_cache-5f4edc62-4320-4aa4-ab91-ee77b757bb7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.101418] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.374997] env[62183]: DEBUG nova.network.neutron [req-d932f48b-62f7-4a78-9259-b5ce0eaa78dd req-0b1a1d21-4086-4d26-a401-46b862979037 service nova] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.451118] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.451689] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 670.456434] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.106s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.458347] env[62183]: INFO nova.compute.claims [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.475872] env[62183]: DEBUG nova.network.neutron [req-d932f48b-62f7-4a78-9259-b5ce0eaa78dd req-0b1a1d21-4086-4d26-a401-46b862979037 service nova] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.957986] env[62183]: DEBUG nova.compute.utils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.959423] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 670.959603] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 670.980724] env[62183]: DEBUG oslo_concurrency.lockutils [req-d932f48b-62f7-4a78-9259-b5ce0eaa78dd req-0b1a1d21-4086-4d26-a401-46b862979037 service nova] Releasing lock "refresh_cache-5f4edc62-4320-4aa4-ab91-ee77b757bb7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.980724] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquired lock "refresh_cache-5f4edc62-4320-4aa4-ab91-ee77b757bb7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.980724] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.004063] env[62183]: DEBUG nova.policy [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70c6875c5ddf4743bb48d8c66ca6c076', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4f739b5563b479eaaef1d5be5dcd847', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 671.298522] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Successfully created port: 213e9902-c275-4eb1-9741-97c81d49d65f {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 671.463081] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 671.510490] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.802826] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.834220] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5659be8-05ec-4ea2-b060-99083aa8158e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.846383] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74748823-b668-4754-8383-5913a4276ef0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.882736] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e405dc79-6c11-4a2a-b039-ac453f92be0d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.891398] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9352dd9f-dd8f-40b4-9240-7140ff40e21a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.906981] env[62183]: DEBUG nova.compute.provider_tree [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.960239] env[62183]: DEBUG nova.compute.manager [req-4a8064d2-a13c-40c2-b67b-c68b675e0f97 req-e6d0384a-c272-4bdf-96fc-fe10ff72e231 service nova] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Received event network-vif-deleted-efcc7c0d-32ee-4834-84b0-a32619083b33 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 672.305763] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Releasing lock "refresh_cache-5f4edc62-4320-4aa4-ab91-ee77b757bb7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.306222] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 672.306682] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 672.306737] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d649fdf-50c2-4dbc-b4f3-99ae0e347159 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.321430] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd86a0b-41dd-4cd0-b814-4b318aede079 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.346160] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5f4edc62-4320-4aa4-ab91-ee77b757bb7c could not be found. [ 672.346160] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 672.346160] env[62183]: INFO nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 672.346160] env[62183]: DEBUG oslo.service.loopingcall [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.346160] env[62183]: DEBUG nova.compute.manager [-] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.346160] env[62183]: DEBUG nova.network.neutron [-] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.375943] env[62183]: DEBUG nova.network.neutron [-] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.413251] env[62183]: DEBUG nova.scheduler.client.report [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 672.484333] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 672.515017] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 672.515017] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 672.515017] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 672.515381] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 672.515381] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 672.515381] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 672.515381] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 672.515381] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 672.515527] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 672.515527] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 672.515527] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.515527] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c264153-d361-494c-97cd-ffa8ee0d3e97 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.525300] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73dd3f7f-fe07-47a1-af59-2dc89d4b2a49 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.559838] env[62183]: ERROR nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 213e9902-c275-4eb1-9741-97c81d49d65f, please check neutron logs for more information. [ 672.559838] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 672.559838] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.559838] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 672.559838] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.559838] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 672.559838] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.559838] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 672.559838] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.559838] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 672.559838] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.559838] env[62183]: ERROR nova.compute.manager raise self.value [ 672.559838] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.559838] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 672.559838] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.559838] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 672.560290] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.560290] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 672.560290] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 213e9902-c275-4eb1-9741-97c81d49d65f, please check neutron logs for more information. [ 672.560290] env[62183]: ERROR nova.compute.manager [ 672.560290] env[62183]: Traceback (most recent call last): [ 672.560290] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 672.560290] env[62183]: listener.cb(fileno) [ 672.560290] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.560290] env[62183]: result = function(*args, **kwargs) [ 672.560290] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.560290] env[62183]: return func(*args, **kwargs) [ 672.560290] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.560290] env[62183]: raise e [ 672.560290] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.560290] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 672.560290] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.560290] env[62183]: created_port_ids = self._update_ports_for_instance( [ 672.560290] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.560290] env[62183]: with excutils.save_and_reraise_exception(): [ 672.560290] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.560290] env[62183]: self.force_reraise() [ 672.560290] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.560290] env[62183]: raise self.value [ 672.560290] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.560290] env[62183]: updated_port = self._update_port( [ 672.560290] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.560290] env[62183]: _ensure_no_port_binding_failure(port) [ 672.560290] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.560290] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 672.560982] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 213e9902-c275-4eb1-9741-97c81d49d65f, please check neutron logs for more information. [ 672.560982] env[62183]: Removing descriptor: 14 [ 672.560982] env[62183]: ERROR nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 213e9902-c275-4eb1-9741-97c81d49d65f, please check neutron logs for more information. [ 672.560982] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Traceback (most recent call last): [ 672.560982] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 672.560982] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] yield resources [ 672.560982] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.560982] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self.driver.spawn(context, instance, image_meta, [ 672.560982] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 672.560982] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.560982] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.560982] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] vm_ref = self.build_virtual_machine(instance, [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] for vif in network_info: [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] return self._sync_wrapper(fn, *args, **kwargs) [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self.wait() [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self[:] = self._gt.wait() [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] return self._exit_event.wait() [ 672.561389] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] result = hub.switch() [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] return self.greenlet.switch() [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] result = function(*args, **kwargs) [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] return func(*args, **kwargs) [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] raise e [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] nwinfo = self.network_api.allocate_for_instance( [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.561712] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] created_port_ids = self._update_ports_for_instance( [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] with excutils.save_and_reraise_exception(): [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self.force_reraise() [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] raise self.value [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] updated_port = self._update_port( [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] _ensure_no_port_binding_failure(port) [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.562117] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] raise exception.PortBindingFailed(port_id=port['id']) [ 672.562419] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] nova.exception.PortBindingFailed: Binding failed for port 213e9902-c275-4eb1-9741-97c81d49d65f, please check neutron logs for more information. [ 672.562419] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] [ 672.562419] env[62183]: INFO nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Terminating instance [ 672.563137] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "refresh_cache-f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.563304] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquired lock "refresh_cache-f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.563470] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.877533] env[62183]: DEBUG nova.network.neutron [-] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.917300] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.917300] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 672.920425] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.566s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.921926] env[62183]: INFO nova.compute.claims [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.082149] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.225184] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.381599] env[62183]: INFO nova.compute.manager [-] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Took 1.04 seconds to deallocate network for instance. [ 673.384465] env[62183]: DEBUG nova.compute.claims [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 673.384650] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.432318] env[62183]: DEBUG nova.compute.utils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.438409] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.438409] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 673.479150] env[62183]: DEBUG nova.policy [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '70c6875c5ddf4743bb48d8c66ca6c076', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4f739b5563b479eaaef1d5be5dcd847', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.729205] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Releasing lock "refresh_cache-f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.730034] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 673.730034] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 673.730209] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4ef065c-996f-4164-b87a-d4d0784f1e49 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.744013] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545d1393-493a-4291-bc04-6bc22fb2f2d2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.766583] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f4b29d0e-488e-4d0c-b2cb-1321c40c87ea could not be found. [ 673.766960] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 673.767220] env[62183]: INFO nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 673.767504] env[62183]: DEBUG oslo.service.loopingcall [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.771052] env[62183]: DEBUG nova.compute.manager [-] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.771052] env[62183]: DEBUG nova.network.neutron [-] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.795468] env[62183]: DEBUG nova.network.neutron [-] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.824584] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Successfully created port: c798064c-381f-4a76-917c-2c5550c0a248 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.940763] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 673.995372] env[62183]: DEBUG nova.compute.manager [req-5643eddd-b3dc-4578-9d51-a5756b584198 req-74462fe1-b527-46f8-88e9-503ee63678dc service nova] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Received event network-changed-213e9902-c275-4eb1-9741-97c81d49d65f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 673.995562] env[62183]: DEBUG nova.compute.manager [req-5643eddd-b3dc-4578-9d51-a5756b584198 req-74462fe1-b527-46f8-88e9-503ee63678dc service nova] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Refreshing instance network info cache due to event network-changed-213e9902-c275-4eb1-9741-97c81d49d65f. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 673.995776] env[62183]: DEBUG oslo_concurrency.lockutils [req-5643eddd-b3dc-4578-9d51-a5756b584198 req-74462fe1-b527-46f8-88e9-503ee63678dc service nova] Acquiring lock "refresh_cache-f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.995918] env[62183]: DEBUG oslo_concurrency.lockutils [req-5643eddd-b3dc-4578-9d51-a5756b584198 req-74462fe1-b527-46f8-88e9-503ee63678dc service nova] Acquired lock "refresh_cache-f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.996094] env[62183]: DEBUG nova.network.neutron [req-5643eddd-b3dc-4578-9d51-a5756b584198 req-74462fe1-b527-46f8-88e9-503ee63678dc service nova] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Refreshing network info cache for port 213e9902-c275-4eb1-9741-97c81d49d65f {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 674.301807] env[62183]: DEBUG nova.network.neutron [-] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.323987] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7adbb2-6552-4413-9c8a-78b816790c79 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.332741] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90739842-b69d-454e-9c6c-1d789d3360f4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.374276] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54a1f55-5f85-4e43-9253-4a6890e9912f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.381937] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612efaa6-5500-462a-b8f9-1def694e33ac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.397713] env[62183]: DEBUG nova.compute.provider_tree [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.525524] env[62183]: DEBUG nova.network.neutron [req-5643eddd-b3dc-4578-9d51-a5756b584198 req-74462fe1-b527-46f8-88e9-503ee63678dc service nova] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.667884] env[62183]: DEBUG nova.network.neutron [req-5643eddd-b3dc-4578-9d51-a5756b584198 req-74462fe1-b527-46f8-88e9-503ee63678dc service nova] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.804481] env[62183]: INFO nova.compute.manager [-] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Took 1.03 seconds to deallocate network for instance. [ 674.806855] env[62183]: DEBUG nova.compute.claims [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 674.807042] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.902561] env[62183]: DEBUG nova.scheduler.client.report [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 674.955630] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 674.983587] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.983881] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.984137] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.984377] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.984547] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.984729] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.984958] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.985181] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.985385] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.985568] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.985737] env[62183]: DEBUG nova.virt.hardware [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.986694] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d152c9-39be-4f59-a42d-d7be766fdb9b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.995318] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bbc49e-2a40-45ea-8775-113daaed28d9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.075394] env[62183]: ERROR nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c798064c-381f-4a76-917c-2c5550c0a248, please check neutron logs for more information. [ 675.075394] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.075394] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.075394] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.075394] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.075394] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.075394] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.075394] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.075394] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.075394] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 675.075394] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.075394] env[62183]: ERROR nova.compute.manager raise self.value [ 675.075394] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.075394] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.075394] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.075394] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.075877] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.075877] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.075877] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c798064c-381f-4a76-917c-2c5550c0a248, please check neutron logs for more information. [ 675.075877] env[62183]: ERROR nova.compute.manager [ 675.075877] env[62183]: Traceback (most recent call last): [ 675.075877] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.075877] env[62183]: listener.cb(fileno) [ 675.075877] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.075877] env[62183]: result = function(*args, **kwargs) [ 675.075877] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 675.075877] env[62183]: return func(*args, **kwargs) [ 675.075877] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.075877] env[62183]: raise e [ 675.075877] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.075877] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 675.075877] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.075877] env[62183]: created_port_ids = self._update_ports_for_instance( [ 675.075877] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.075877] env[62183]: with excutils.save_and_reraise_exception(): [ 675.075877] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.075877] env[62183]: self.force_reraise() [ 675.075877] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.075877] env[62183]: raise self.value [ 675.075877] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.075877] env[62183]: updated_port = self._update_port( [ 675.075877] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.075877] env[62183]: _ensure_no_port_binding_failure(port) [ 675.075877] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.075877] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.076662] env[62183]: nova.exception.PortBindingFailed: Binding failed for port c798064c-381f-4a76-917c-2c5550c0a248, please check neutron logs for more information. [ 675.076662] env[62183]: Removing descriptor: 14 [ 675.076662] env[62183]: ERROR nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c798064c-381f-4a76-917c-2c5550c0a248, please check neutron logs for more information. [ 675.076662] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Traceback (most recent call last): [ 675.076662] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 675.076662] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] yield resources [ 675.076662] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.076662] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self.driver.spawn(context, instance, image_meta, [ 675.076662] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 675.076662] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.076662] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.076662] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] vm_ref = self.build_virtual_machine(instance, [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] for vif in network_info: [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] return self._sync_wrapper(fn, *args, **kwargs) [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self.wait() [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self[:] = self._gt.wait() [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] return self._exit_event.wait() [ 675.077038] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] result = hub.switch() [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] return self.greenlet.switch() [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] result = function(*args, **kwargs) [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] return func(*args, **kwargs) [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] raise e [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] nwinfo = self.network_api.allocate_for_instance( [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.077412] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] created_port_ids = self._update_ports_for_instance( [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] with excutils.save_and_reraise_exception(): [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self.force_reraise() [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] raise self.value [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] updated_port = self._update_port( [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] _ensure_no_port_binding_failure(port) [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.077756] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] raise exception.PortBindingFailed(port_id=port['id']) [ 675.078115] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] nova.exception.PortBindingFailed: Binding failed for port c798064c-381f-4a76-917c-2c5550c0a248, please check neutron logs for more information. [ 675.078115] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] [ 675.078115] env[62183]: INFO nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Terminating instance [ 675.078750] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "refresh_cache-0120298f-dc4c-4c13-ab71-70ccd6428dc9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.078911] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquired lock "refresh_cache-0120298f-dc4c-4c13-ab71-70ccd6428dc9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.079085] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.171031] env[62183]: DEBUG oslo_concurrency.lockutils [req-5643eddd-b3dc-4578-9d51-a5756b584198 req-74462fe1-b527-46f8-88e9-503ee63678dc service nova] Releasing lock "refresh_cache-f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.171300] env[62183]: DEBUG nova.compute.manager [req-5643eddd-b3dc-4578-9d51-a5756b584198 req-74462fe1-b527-46f8-88e9-503ee63678dc service nova] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Received event network-vif-deleted-213e9902-c275-4eb1-9741-97c81d49d65f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.416382] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.416918] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.422533] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.838s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.598899] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.680518] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.924598] env[62183]: DEBUG nova.compute.utils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.929422] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.929597] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 676.042963] env[62183]: DEBUG nova.compute.manager [req-6df19665-01db-4297-906a-4c73493726f7 req-ec3e7bb2-0558-4fa9-83d6-ee7934cd00b8 service nova] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Received event network-changed-c798064c-381f-4a76-917c-2c5550c0a248 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 676.043194] env[62183]: DEBUG nova.compute.manager [req-6df19665-01db-4297-906a-4c73493726f7 req-ec3e7bb2-0558-4fa9-83d6-ee7934cd00b8 service nova] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Refreshing instance network info cache due to event network-changed-c798064c-381f-4a76-917c-2c5550c0a248. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 676.044304] env[62183]: DEBUG oslo_concurrency.lockutils [req-6df19665-01db-4297-906a-4c73493726f7 req-ec3e7bb2-0558-4fa9-83d6-ee7934cd00b8 service nova] Acquiring lock "refresh_cache-0120298f-dc4c-4c13-ab71-70ccd6428dc9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.054653] env[62183]: DEBUG nova.policy [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9657e30b5c04e4d95f0ab638d411469', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa3331f62f874b1d90b318add34b3786', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 676.184750] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Releasing lock "refresh_cache-0120298f-dc4c-4c13-ab71-70ccd6428dc9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.185359] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 676.185581] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 676.186090] env[62183]: DEBUG oslo_concurrency.lockutils [req-6df19665-01db-4297-906a-4c73493726f7 req-ec3e7bb2-0558-4fa9-83d6-ee7934cd00b8 service nova] Acquired lock "refresh_cache-0120298f-dc4c-4c13-ab71-70ccd6428dc9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.187339] env[62183]: DEBUG nova.network.neutron [req-6df19665-01db-4297-906a-4c73493726f7 req-ec3e7bb2-0558-4fa9-83d6-ee7934cd00b8 service nova] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Refreshing network info cache for port c798064c-381f-4a76-917c-2c5550c0a248 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 676.187339] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f6673a4c-c490-495b-a56b-b29b9aa57ff2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.198543] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899150d4-d788-4940-8ca8-6a5eccbfe4e9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.226991] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0120298f-dc4c-4c13-ab71-70ccd6428dc9 could not be found. [ 676.227244] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.227427] env[62183]: INFO nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 676.227874] env[62183]: DEBUG oslo.service.loopingcall [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.230544] env[62183]: DEBUG nova.compute.manager [-] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.230664] env[62183]: DEBUG nova.network.neutron [-] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.255354] env[62183]: DEBUG nova.network.neutron [-] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.308773] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f793c69-ef18-40db-8ff7-9d46b104998e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.315952] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7868c3-ddbf-4f09-9a89-5bbd1a540270 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.349984] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9c3d54-f9de-4d43-b1a6-00f82634b53f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.358165] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a06e0b-6ee7-4462-a2a7-3929df074828 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.371140] env[62183]: DEBUG nova.compute.provider_tree [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.430139] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.631242] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Successfully created port: 91502b7d-747c-46a9-9590-5442cc87cf14 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.710563] env[62183]: DEBUG nova.network.neutron [req-6df19665-01db-4297-906a-4c73493726f7 req-ec3e7bb2-0558-4fa9-83d6-ee7934cd00b8 service nova] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.759548] env[62183]: DEBUG nova.network.neutron [-] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.864767] env[62183]: DEBUG nova.network.neutron [req-6df19665-01db-4297-906a-4c73493726f7 req-ec3e7bb2-0558-4fa9-83d6-ee7934cd00b8 service nova] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.874925] env[62183]: DEBUG nova.scheduler.client.report [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 676.999284] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Successfully created port: 9b5d61c6-7d99-4f6c-a8c1-dbe660af350c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 677.264108] env[62183]: INFO nova.compute.manager [-] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Took 1.03 seconds to deallocate network for instance. [ 677.266440] env[62183]: DEBUG nova.compute.claims [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 677.266615] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.367874] env[62183]: DEBUG oslo_concurrency.lockutils [req-6df19665-01db-4297-906a-4c73493726f7 req-ec3e7bb2-0558-4fa9-83d6-ee7934cd00b8 service nova] Releasing lock "refresh_cache-0120298f-dc4c-4c13-ab71-70ccd6428dc9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.367962] env[62183]: DEBUG nova.compute.manager [req-6df19665-01db-4297-906a-4c73493726f7 req-ec3e7bb2-0558-4fa9-83d6-ee7934cd00b8 service nova] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Received event network-vif-deleted-c798064c-381f-4a76-917c-2c5550c0a248 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.379148] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.379803] env[62183]: ERROR nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port abc2f404-a0c3-4865-b54f-701e0a25728d, please check neutron logs for more information. [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Traceback (most recent call last): [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self.driver.spawn(context, instance, image_meta, [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] vm_ref = self.build_virtual_machine(instance, [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.379803] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] for vif in network_info: [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] return self._sync_wrapper(fn, *args, **kwargs) [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self.wait() [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self[:] = self._gt.wait() [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] return self._exit_event.wait() [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] result = hub.switch() [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.380194] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] return self.greenlet.switch() [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] result = function(*args, **kwargs) [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] return func(*args, **kwargs) [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] raise e [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] nwinfo = self.network_api.allocate_for_instance( [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] created_port_ids = self._update_ports_for_instance( [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] with excutils.save_and_reraise_exception(): [ 677.380565] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] self.force_reraise() [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] raise self.value [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] updated_port = self._update_port( [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] _ensure_no_port_binding_failure(port) [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] raise exception.PortBindingFailed(port_id=port['id']) [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] nova.exception.PortBindingFailed: Binding failed for port abc2f404-a0c3-4865-b54f-701e0a25728d, please check neutron logs for more information. [ 677.380965] env[62183]: ERROR nova.compute.manager [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] [ 677.381291] env[62183]: DEBUG nova.compute.utils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Binding failed for port abc2f404-a0c3-4865-b54f-701e0a25728d, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.381923] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.223s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.383543] env[62183]: INFO nova.compute.claims [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.388655] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Build of instance 8e7fbc90-5be0-4baf-b371-1a22e84dee98 was re-scheduled: Binding failed for port abc2f404-a0c3-4865-b54f-701e0a25728d, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.388655] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.388655] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquiring lock "refresh_cache-8e7fbc90-5be0-4baf-b371-1a22e84dee98" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.388655] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Acquired lock "refresh_cache-8e7fbc90-5be0-4baf-b371-1a22e84dee98" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.388927] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.444549] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.476103] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.476336] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.476488] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.476761] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.476819] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.476944] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.477162] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.477320] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.477484] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.477642] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.477810] env[62183]: DEBUG nova.virt.hardware [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.478683] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09066164-0cc3-4625-a83a-856fbac4d55b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.486335] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38583e2d-6ced-42e6-81ff-da2dd04fd566 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.910822] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.052086] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.085880] env[62183]: DEBUG nova.compute.manager [req-e3f5007d-8f04-40e5-9888-eb7b03ddcfe8 req-289f5fba-18d2-43e1-89cb-372845cf1855 service nova] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Received event network-changed-91502b7d-747c-46a9-9590-5442cc87cf14 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 678.086083] env[62183]: DEBUG nova.compute.manager [req-e3f5007d-8f04-40e5-9888-eb7b03ddcfe8 req-289f5fba-18d2-43e1-89cb-372845cf1855 service nova] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Refreshing instance network info cache due to event network-changed-91502b7d-747c-46a9-9590-5442cc87cf14. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 678.086294] env[62183]: DEBUG oslo_concurrency.lockutils [req-e3f5007d-8f04-40e5-9888-eb7b03ddcfe8 req-289f5fba-18d2-43e1-89cb-372845cf1855 service nova] Acquiring lock "refresh_cache-9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.086494] env[62183]: DEBUG oslo_concurrency.lockutils [req-e3f5007d-8f04-40e5-9888-eb7b03ddcfe8 req-289f5fba-18d2-43e1-89cb-372845cf1855 service nova] Acquired lock "refresh_cache-9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.086575] env[62183]: DEBUG nova.network.neutron [req-e3f5007d-8f04-40e5-9888-eb7b03ddcfe8 req-289f5fba-18d2-43e1-89cb-372845cf1855 service nova] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Refreshing network info cache for port 91502b7d-747c-46a9-9590-5442cc87cf14 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 678.268399] env[62183]: ERROR nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 91502b7d-747c-46a9-9590-5442cc87cf14, please check neutron logs for more information. [ 678.268399] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 678.268399] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.268399] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 678.268399] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.268399] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 678.268399] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.268399] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 678.268399] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.268399] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 678.268399] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.268399] env[62183]: ERROR nova.compute.manager raise self.value [ 678.268399] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.268399] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 678.268399] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.268399] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 678.268889] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.268889] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 678.268889] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 91502b7d-747c-46a9-9590-5442cc87cf14, please check neutron logs for more information. [ 678.268889] env[62183]: ERROR nova.compute.manager [ 678.268889] env[62183]: Traceback (most recent call last): [ 678.268889] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 678.268889] env[62183]: listener.cb(fileno) [ 678.268889] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.268889] env[62183]: result = function(*args, **kwargs) [ 678.268889] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 678.268889] env[62183]: return func(*args, **kwargs) [ 678.268889] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.268889] env[62183]: raise e [ 678.268889] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.268889] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 678.268889] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.268889] env[62183]: created_port_ids = self._update_ports_for_instance( [ 678.268889] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.268889] env[62183]: with excutils.save_and_reraise_exception(): [ 678.268889] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.268889] env[62183]: self.force_reraise() [ 678.268889] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.268889] env[62183]: raise self.value [ 678.268889] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.268889] env[62183]: updated_port = self._update_port( [ 678.268889] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.268889] env[62183]: _ensure_no_port_binding_failure(port) [ 678.268889] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.268889] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 678.269759] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 91502b7d-747c-46a9-9590-5442cc87cf14, please check neutron logs for more information. [ 678.269759] env[62183]: Removing descriptor: 14 [ 678.269759] env[62183]: ERROR nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 91502b7d-747c-46a9-9590-5442cc87cf14, please check neutron logs for more information. [ 678.269759] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Traceback (most recent call last): [ 678.269759] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 678.269759] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] yield resources [ 678.269759] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 678.269759] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self.driver.spawn(context, instance, image_meta, [ 678.269759] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 678.269759] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 678.269759] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 678.269759] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] vm_ref = self.build_virtual_machine(instance, [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] for vif in network_info: [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] return self._sync_wrapper(fn, *args, **kwargs) [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self.wait() [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self[:] = self._gt.wait() [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] return self._exit_event.wait() [ 678.270208] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] result = hub.switch() [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] return self.greenlet.switch() [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] result = function(*args, **kwargs) [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] return func(*args, **kwargs) [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] raise e [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] nwinfo = self.network_api.allocate_for_instance( [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 678.270607] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] created_port_ids = self._update_ports_for_instance( [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] with excutils.save_and_reraise_exception(): [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self.force_reraise() [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] raise self.value [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] updated_port = self._update_port( [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] _ensure_no_port_binding_failure(port) [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 678.271071] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] raise exception.PortBindingFailed(port_id=port['id']) [ 678.271460] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] nova.exception.PortBindingFailed: Binding failed for port 91502b7d-747c-46a9-9590-5442cc87cf14, please check neutron logs for more information. [ 678.271460] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] [ 678.271460] env[62183]: INFO nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Terminating instance [ 678.272347] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "refresh_cache-9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.555863] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Releasing lock "refresh_cache-8e7fbc90-5be0-4baf-b371-1a22e84dee98" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.556118] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.556307] env[62183]: DEBUG nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.556468] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.572711] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.624543] env[62183]: DEBUG nova.network.neutron [req-e3f5007d-8f04-40e5-9888-eb7b03ddcfe8 req-289f5fba-18d2-43e1-89cb-372845cf1855 service nova] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.689144] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e60e7a-2418-4b8c-ab99-aa3c364d26a8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.696870] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b23e62b-44b6-4101-bb79-cc3c0f865728 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.726292] env[62183]: DEBUG nova.network.neutron [req-e3f5007d-8f04-40e5-9888-eb7b03ddcfe8 req-289f5fba-18d2-43e1-89cb-372845cf1855 service nova] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.727828] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46763161-5965-4e60-9c46-30ecff6a9eae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.735081] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f3f373-9582-4192-bcb8-39f03f8b802e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.749882] env[62183]: DEBUG nova.compute.provider_tree [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.075230] env[62183]: DEBUG nova.network.neutron [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.231428] env[62183]: DEBUG oslo_concurrency.lockutils [req-e3f5007d-8f04-40e5-9888-eb7b03ddcfe8 req-289f5fba-18d2-43e1-89cb-372845cf1855 service nova] Releasing lock "refresh_cache-9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.232381] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquired lock "refresh_cache-9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.232381] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 679.253179] env[62183]: DEBUG nova.scheduler.client.report [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 679.577879] env[62183]: INFO nova.compute.manager [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] [instance: 8e7fbc90-5be0-4baf-b371-1a22e84dee98] Took 1.02 seconds to deallocate network for instance. [ 679.758714] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.759449] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 679.763316] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.684s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.765760] env[62183]: INFO nova.compute.claims [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.770389] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.898721] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.112422] env[62183]: DEBUG nova.compute.manager [req-8fb3919c-a542-4c8b-9647-5d70abf3f750 req-bf388cfa-42ac-4ea1-9f65-bb2a55d1556e service nova] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Received event network-vif-deleted-91502b7d-747c-46a9-9590-5442cc87cf14 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.273917] env[62183]: DEBUG nova.compute.utils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.276916] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 680.277100] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 680.325765] env[62183]: DEBUG nova.policy [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'acb900da57a64e358fe3fe87ea1bcb72', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66914c33ade24f9a802f3731240205be', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 680.402907] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Releasing lock "refresh_cache-9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.402907] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 680.402907] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 680.402907] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89a80371-2d6e-4579-bcb9-6e403187b323 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.412140] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fd0065-adfa-45cb-8cd2-dce2e3169dca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.434577] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4 could not be found. [ 680.434794] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 680.434976] env[62183]: INFO nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 680.435231] env[62183]: DEBUG oslo.service.loopingcall [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 680.435450] env[62183]: DEBUG nova.compute.manager [-] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.435581] env[62183]: DEBUG nova.network.neutron [-] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 680.476315] env[62183]: DEBUG nova.network.neutron [-] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.616028] env[62183]: INFO nova.scheduler.client.report [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Deleted allocations for instance 8e7fbc90-5be0-4baf-b371-1a22e84dee98 [ 680.728791] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Successfully created port: 5926ff4f-0602-40e3-8704-b0dcb398089d {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.783902] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 681.106110] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597a4f3c-182c-4469-98f2-4db53190153d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.116028] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401cf323-a36a-4be0-b99d-d2b52243a927 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.143821] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c77ef97d-c3da-471f-a33a-7361c9fb373c tempest-VolumesAdminNegativeTest-2126586946 tempest-VolumesAdminNegativeTest-2126586946-project-member] Lock "8e7fbc90-5be0-4baf-b371-1a22e84dee98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.808s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.147111] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1054792c-b903-4d43-9659-694dfb40c9bd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.154503] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4649b5dc-0537-411a-9afc-4b30c441511f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.168064] env[62183]: DEBUG nova.compute.provider_tree [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.291520] env[62183]: DEBUG nova.network.neutron [-] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.650128] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.670979] env[62183]: DEBUG nova.scheduler.client.report [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 681.794664] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 681.798376] env[62183]: INFO nova.compute.manager [-] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Took 1.36 seconds to deallocate network for instance. [ 681.801281] env[62183]: DEBUG nova.compute.claims [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 681.801471] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.824907] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.825155] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.825416] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.825522] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.825623] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.825766] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.825967] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.826139] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.826403] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.826486] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.826608] env[62183]: DEBUG nova.virt.hardware [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.827744] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8175eb0d-7843-41c2-8696-3442805f9bd4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.835481] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f437d5c-d877-4132-9afb-9976f11b1f6b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.944076] env[62183]: ERROR nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5926ff4f-0602-40e3-8704-b0dcb398089d, please check neutron logs for more information. [ 681.944076] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 681.944076] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.944076] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 681.944076] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.944076] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 681.944076] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.944076] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 681.944076] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.944076] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 681.944076] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.944076] env[62183]: ERROR nova.compute.manager raise self.value [ 681.944076] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.944076] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 681.944076] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.944076] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 681.944503] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.944503] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 681.944503] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5926ff4f-0602-40e3-8704-b0dcb398089d, please check neutron logs for more information. [ 681.944503] env[62183]: ERROR nova.compute.manager [ 681.944503] env[62183]: Traceback (most recent call last): [ 681.944503] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 681.944503] env[62183]: listener.cb(fileno) [ 681.944503] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.944503] env[62183]: result = function(*args, **kwargs) [ 681.944503] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.944503] env[62183]: return func(*args, **kwargs) [ 681.944503] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.944503] env[62183]: raise e [ 681.944503] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.944503] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 681.944503] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.944503] env[62183]: created_port_ids = self._update_ports_for_instance( [ 681.944503] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.944503] env[62183]: with excutils.save_and_reraise_exception(): [ 681.944503] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.944503] env[62183]: self.force_reraise() [ 681.944503] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.944503] env[62183]: raise self.value [ 681.944503] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.944503] env[62183]: updated_port = self._update_port( [ 681.944503] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.944503] env[62183]: _ensure_no_port_binding_failure(port) [ 681.944503] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.944503] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 681.945274] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 5926ff4f-0602-40e3-8704-b0dcb398089d, please check neutron logs for more information. [ 681.945274] env[62183]: Removing descriptor: 14 [ 681.945274] env[62183]: ERROR nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5926ff4f-0602-40e3-8704-b0dcb398089d, please check neutron logs for more information. [ 681.945274] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Traceback (most recent call last): [ 681.945274] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 681.945274] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] yield resources [ 681.945274] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.945274] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self.driver.spawn(context, instance, image_meta, [ 681.945274] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 681.945274] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.945274] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.945274] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] vm_ref = self.build_virtual_machine(instance, [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] for vif in network_info: [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] return self._sync_wrapper(fn, *args, **kwargs) [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self.wait() [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self[:] = self._gt.wait() [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] return self._exit_event.wait() [ 681.945607] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] result = hub.switch() [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] return self.greenlet.switch() [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] result = function(*args, **kwargs) [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] return func(*args, **kwargs) [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] raise e [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] nwinfo = self.network_api.allocate_for_instance( [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.945931] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] created_port_ids = self._update_ports_for_instance( [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] with excutils.save_and_reraise_exception(): [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self.force_reraise() [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] raise self.value [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] updated_port = self._update_port( [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] _ensure_no_port_binding_failure(port) [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.946835] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] raise exception.PortBindingFailed(port_id=port['id']) [ 681.947321] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] nova.exception.PortBindingFailed: Binding failed for port 5926ff4f-0602-40e3-8704-b0dcb398089d, please check neutron logs for more information. [ 681.947321] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] [ 681.947321] env[62183]: INFO nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Terminating instance [ 681.948128] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Acquiring lock "refresh_cache-f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.948128] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Acquired lock "refresh_cache-f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.948128] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.139235] env[62183]: DEBUG nova.compute.manager [req-2fe1f3c9-544d-4c12-b498-5deb8ef608fd req-b16efe11-2902-4d6b-a796-fdfb4f579c7d service nova] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Received event network-changed-5926ff4f-0602-40e3-8704-b0dcb398089d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.139500] env[62183]: DEBUG nova.compute.manager [req-2fe1f3c9-544d-4c12-b498-5deb8ef608fd req-b16efe11-2902-4d6b-a796-fdfb4f579c7d service nova] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Refreshing instance network info cache due to event network-changed-5926ff4f-0602-40e3-8704-b0dcb398089d. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 682.139575] env[62183]: DEBUG oslo_concurrency.lockutils [req-2fe1f3c9-544d-4c12-b498-5deb8ef608fd req-b16efe11-2902-4d6b-a796-fdfb4f579c7d service nova] Acquiring lock "refresh_cache-f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.176171] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.176687] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 682.180126] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.180403] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.075s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.181999] env[62183]: INFO nova.compute.claims [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.475666] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.642906] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.686949] env[62183]: DEBUG nova.compute.utils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.690187] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.690374] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 682.747608] env[62183]: DEBUG nova.policy [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89ad04efd23b40e6a3eea5b3e6fd021e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b61c6a99cbda435481bb72f20929f03f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 683.145816] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Releasing lock "refresh_cache-f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.146279] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.146487] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.146829] env[62183]: DEBUG oslo_concurrency.lockutils [req-2fe1f3c9-544d-4c12-b498-5deb8ef608fd req-b16efe11-2902-4d6b-a796-fdfb4f579c7d service nova] Acquired lock "refresh_cache-f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.147014] env[62183]: DEBUG nova.network.neutron [req-2fe1f3c9-544d-4c12-b498-5deb8ef608fd req-b16efe11-2902-4d6b-a796-fdfb4f579c7d service nova] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Refreshing network info cache for port 5926ff4f-0602-40e3-8704-b0dcb398089d {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 683.148127] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-765cd0ea-8262-4b1e-97d7-5424b5dd8d11 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.160047] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767f403f-06bb-4ed1-b0cd-72302fb5c84a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.186580] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f67ab6cd-410b-4ed3-890b-68f3b0f41ec7 could not be found. [ 683.186827] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.187014] env[62183]: INFO nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 683.187263] env[62183]: DEBUG oslo.service.loopingcall [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.187488] env[62183]: DEBUG nova.compute.manager [-] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.187658] env[62183]: DEBUG nova.network.neutron [-] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.190901] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 683.220904] env[62183]: DEBUG nova.network.neutron [-] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.365197] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Successfully created port: 9128c251-e7f0-4964-8590-e1cae8fe8f96 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.520265] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64bbfc92-6cce-4049-bbf4-6f0940576140 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.527806] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438d2a08-8039-484d-abb8-29d5ae0b3947 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.558471] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bac66d5-1715-442e-8372-973710128709 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.565977] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78fab28a-d2db-4714-925f-ebdaee5398b3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.581199] env[62183]: DEBUG nova.compute.provider_tree [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.669154] env[62183]: DEBUG nova.network.neutron [req-2fe1f3c9-544d-4c12-b498-5deb8ef608fd req-b16efe11-2902-4d6b-a796-fdfb4f579c7d service nova] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.723635] env[62183]: DEBUG nova.network.neutron [-] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.796609] env[62183]: DEBUG nova.network.neutron [req-2fe1f3c9-544d-4c12-b498-5deb8ef608fd req-b16efe11-2902-4d6b-a796-fdfb4f579c7d service nova] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.086280] env[62183]: DEBUG nova.scheduler.client.report [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 684.205325] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 684.227367] env[62183]: INFO nova.compute.manager [-] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Took 1.04 seconds to deallocate network for instance. [ 684.233428] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.233756] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.233826] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.233986] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.234145] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.234286] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.234619] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.234681] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.234795] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.234951] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.235546] env[62183]: DEBUG nova.virt.hardware [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.236576] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d221c1a-6ac9-4a75-b497-aa3e97c52c40 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.242419] env[62183]: DEBUG nova.compute.claims [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.242662] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.243708] env[62183]: DEBUG nova.compute.manager [req-ac92f1ed-524a-41ea-be71-4030076e3cef req-0e79408b-57c4-4aad-bc02-730e0ed5bced service nova] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Received event network-changed-9128c251-e7f0-4964-8590-e1cae8fe8f96 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.243953] env[62183]: DEBUG nova.compute.manager [req-ac92f1ed-524a-41ea-be71-4030076e3cef req-0e79408b-57c4-4aad-bc02-730e0ed5bced service nova] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Refreshing instance network info cache due to event network-changed-9128c251-e7f0-4964-8590-e1cae8fe8f96. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 684.244491] env[62183]: DEBUG oslo_concurrency.lockutils [req-ac92f1ed-524a-41ea-be71-4030076e3cef req-0e79408b-57c4-4aad-bc02-730e0ed5bced service nova] Acquiring lock "refresh_cache-ee99cd66-908e-448f-9cce-ebe78a32214d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.244491] env[62183]: DEBUG oslo_concurrency.lockutils [req-ac92f1ed-524a-41ea-be71-4030076e3cef req-0e79408b-57c4-4aad-bc02-730e0ed5bced service nova] Acquired lock "refresh_cache-ee99cd66-908e-448f-9cce-ebe78a32214d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.244604] env[62183]: DEBUG nova.network.neutron [req-ac92f1ed-524a-41ea-be71-4030076e3cef req-0e79408b-57c4-4aad-bc02-730e0ed5bced service nova] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Refreshing network info cache for port 9128c251-e7f0-4964-8590-e1cae8fe8f96 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 684.252536] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9112cb47-e1a2-4863-913e-b349ce11b3b3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.300360] env[62183]: DEBUG oslo_concurrency.lockutils [req-2fe1f3c9-544d-4c12-b498-5deb8ef608fd req-b16efe11-2902-4d6b-a796-fdfb4f579c7d service nova] Releasing lock "refresh_cache-f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.300458] env[62183]: DEBUG nova.compute.manager [req-2fe1f3c9-544d-4c12-b498-5deb8ef608fd req-b16efe11-2902-4d6b-a796-fdfb4f579c7d service nova] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Received event network-vif-deleted-5926ff4f-0602-40e3-8704-b0dcb398089d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.451079] env[62183]: ERROR nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9128c251-e7f0-4964-8590-e1cae8fe8f96, please check neutron logs for more information. [ 684.451079] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 684.451079] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.451079] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 684.451079] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.451079] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 684.451079] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.451079] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 684.451079] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.451079] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 684.451079] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.451079] env[62183]: ERROR nova.compute.manager raise self.value [ 684.451079] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.451079] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 684.451079] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.451079] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 684.451569] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.451569] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 684.451569] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9128c251-e7f0-4964-8590-e1cae8fe8f96, please check neutron logs for more information. [ 684.451569] env[62183]: ERROR nova.compute.manager [ 684.451569] env[62183]: Traceback (most recent call last): [ 684.451569] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 684.451569] env[62183]: listener.cb(fileno) [ 684.451569] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.451569] env[62183]: result = function(*args, **kwargs) [ 684.451569] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.451569] env[62183]: return func(*args, **kwargs) [ 684.451569] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.451569] env[62183]: raise e [ 684.451569] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.451569] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 684.451569] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.451569] env[62183]: created_port_ids = self._update_ports_for_instance( [ 684.451569] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.451569] env[62183]: with excutils.save_and_reraise_exception(): [ 684.451569] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.451569] env[62183]: self.force_reraise() [ 684.451569] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.451569] env[62183]: raise self.value [ 684.451569] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.451569] env[62183]: updated_port = self._update_port( [ 684.451569] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.451569] env[62183]: _ensure_no_port_binding_failure(port) [ 684.451569] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.451569] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 684.452465] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 9128c251-e7f0-4964-8590-e1cae8fe8f96, please check neutron logs for more information. [ 684.452465] env[62183]: Removing descriptor: 14 [ 684.452465] env[62183]: ERROR nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9128c251-e7f0-4964-8590-e1cae8fe8f96, please check neutron logs for more information. [ 684.452465] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Traceback (most recent call last): [ 684.452465] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 684.452465] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] yield resources [ 684.452465] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.452465] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self.driver.spawn(context, instance, image_meta, [ 684.452465] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 684.452465] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.452465] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.452465] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] vm_ref = self.build_virtual_machine(instance, [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] for vif in network_info: [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] return self._sync_wrapper(fn, *args, **kwargs) [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self.wait() [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self[:] = self._gt.wait() [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] return self._exit_event.wait() [ 684.452830] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] result = hub.switch() [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] return self.greenlet.switch() [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] result = function(*args, **kwargs) [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] return func(*args, **kwargs) [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] raise e [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] nwinfo = self.network_api.allocate_for_instance( [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.453271] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] created_port_ids = self._update_ports_for_instance( [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] with excutils.save_and_reraise_exception(): [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self.force_reraise() [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] raise self.value [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] updated_port = self._update_port( [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] _ensure_no_port_binding_failure(port) [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.453620] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] raise exception.PortBindingFailed(port_id=port['id']) [ 684.453965] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] nova.exception.PortBindingFailed: Binding failed for port 9128c251-e7f0-4964-8590-e1cae8fe8f96, please check neutron logs for more information. [ 684.453965] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] [ 684.453965] env[62183]: INFO nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Terminating instance [ 684.455028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-ee99cd66-908e-448f-9cce-ebe78a32214d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.595061] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.595061] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 684.597242] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.539s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.599657] env[62183]: INFO nova.compute.claims [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.763077] env[62183]: DEBUG nova.network.neutron [req-ac92f1ed-524a-41ea-be71-4030076e3cef req-0e79408b-57c4-4aad-bc02-730e0ed5bced service nova] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.848608] env[62183]: DEBUG nova.network.neutron [req-ac92f1ed-524a-41ea-be71-4030076e3cef req-0e79408b-57c4-4aad-bc02-730e0ed5bced service nova] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.104016] env[62183]: DEBUG nova.compute.utils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.110755] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 685.111329] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 685.168268] env[62183]: DEBUG nova.policy [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5694c87b3f9b43de81752befb682d11e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c9c5a819b3041918e10c99176946c05', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.357000] env[62183]: DEBUG oslo_concurrency.lockutils [req-ac92f1ed-524a-41ea-be71-4030076e3cef req-0e79408b-57c4-4aad-bc02-730e0ed5bced service nova] Releasing lock "refresh_cache-ee99cd66-908e-448f-9cce-ebe78a32214d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.357000] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-ee99cd66-908e-448f-9cce-ebe78a32214d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.357000] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.498172] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Successfully created port: 11ed88e3-b166-410c-97a4-6d33a650b04c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 685.611216] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 685.877869] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.954887] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a350283-7412-4eea-84a2-527079817894 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.963263] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9cba13d-fae7-465e-8fa2-e04ca1ec4fbe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.995953] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.997506] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3007ebff-9039-494a-901f-8a40dc9e4d34 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.005710] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fab21a1-3998-4d1f-9228-413f26911bfb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.022508] env[62183]: DEBUG nova.compute.provider_tree [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.261040] env[62183]: DEBUG nova.compute.manager [req-08922d47-a86f-40e9-8b10-0c132c33e029 req-b88ddefa-cf20-4c5e-b1f4-86d3ce138571 service nova] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Received event network-vif-deleted-9128c251-e7f0-4964-8590-e1cae8fe8f96 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 686.501826] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-ee99cd66-908e-448f-9cce-ebe78a32214d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.501919] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 686.502102] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 686.502413] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-382d705d-c884-41d1-9293-118748efa1a5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.511829] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0544b59-3438-40f6-9ba0-9d327baecfc1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.525124] env[62183]: DEBUG nova.scheduler.client.report [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 686.538903] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ee99cd66-908e-448f-9cce-ebe78a32214d could not be found. [ 686.539160] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 686.539380] env[62183]: INFO nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 686.543610] env[62183]: DEBUG oslo.service.loopingcall [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.543750] env[62183]: DEBUG nova.compute.manager [-] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.543844] env[62183]: DEBUG nova.network.neutron [-] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 686.546887] env[62183]: ERROR nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 11ed88e3-b166-410c-97a4-6d33a650b04c, please check neutron logs for more information. [ 686.546887] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 686.546887] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.546887] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 686.546887] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.546887] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 686.546887] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.546887] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 686.546887] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.546887] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 686.546887] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.546887] env[62183]: ERROR nova.compute.manager raise self.value [ 686.546887] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.546887] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 686.546887] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.546887] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 686.547594] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.547594] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 686.547594] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 11ed88e3-b166-410c-97a4-6d33a650b04c, please check neutron logs for more information. [ 686.547594] env[62183]: ERROR nova.compute.manager [ 686.547594] env[62183]: Traceback (most recent call last): [ 686.547594] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 686.547594] env[62183]: listener.cb(fileno) [ 686.547594] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.547594] env[62183]: result = function(*args, **kwargs) [ 686.547594] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.547594] env[62183]: return func(*args, **kwargs) [ 686.547594] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.547594] env[62183]: raise e [ 686.547594] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.547594] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 686.547594] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.547594] env[62183]: created_port_ids = self._update_ports_for_instance( [ 686.547594] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.547594] env[62183]: with excutils.save_and_reraise_exception(): [ 686.547594] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.547594] env[62183]: self.force_reraise() [ 686.547594] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.547594] env[62183]: raise self.value [ 686.547594] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.547594] env[62183]: updated_port = self._update_port( [ 686.547594] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.547594] env[62183]: _ensure_no_port_binding_failure(port) [ 686.547594] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.547594] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 686.548927] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 11ed88e3-b166-410c-97a4-6d33a650b04c, please check neutron logs for more information. [ 686.548927] env[62183]: Removing descriptor: 14 [ 686.558158] env[62183]: DEBUG nova.network.neutron [-] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.624796] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 686.652559] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.652812] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.652929] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.653121] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.653267] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.653445] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.653659] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.653815] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.653979] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.654157] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.654329] env[62183]: DEBUG nova.virt.hardware [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.655265] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a13aa60-8c63-4b6f-a858-7583c8f76a0a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.665224] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0918d146-d499-4de1-af76-7f6d6c258dea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.681779] env[62183]: ERROR nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 11ed88e3-b166-410c-97a4-6d33a650b04c, please check neutron logs for more information. [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Traceback (most recent call last): [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] yield resources [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self.driver.spawn(context, instance, image_meta, [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] vm_ref = self.build_virtual_machine(instance, [ 686.681779] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] for vif in network_info: [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] return self._sync_wrapper(fn, *args, **kwargs) [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self.wait() [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self[:] = self._gt.wait() [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] return self._exit_event.wait() [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 686.682129] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] current.throw(*self._exc) [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] result = function(*args, **kwargs) [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] return func(*args, **kwargs) [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] raise e [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] nwinfo = self.network_api.allocate_for_instance( [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] created_port_ids = self._update_ports_for_instance( [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] with excutils.save_and_reraise_exception(): [ 686.682508] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self.force_reraise() [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] raise self.value [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] updated_port = self._update_port( [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] _ensure_no_port_binding_failure(port) [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] raise exception.PortBindingFailed(port_id=port['id']) [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] nova.exception.PortBindingFailed: Binding failed for port 11ed88e3-b166-410c-97a4-6d33a650b04c, please check neutron logs for more information. [ 686.682817] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] [ 686.682817] env[62183]: INFO nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Terminating instance [ 686.684771] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Acquiring lock "refresh_cache-14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.684771] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Acquired lock "refresh_cache-14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.684771] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.045237] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.045764] env[62183]: DEBUG nova.compute.manager [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.048589] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.947s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.052158] env[62183]: INFO nova.compute.claims [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.060547] env[62183]: DEBUG nova.network.neutron [-] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.204711] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.343343] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.554565] env[62183]: DEBUG nova.compute.utils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.556439] env[62183]: DEBUG nova.compute.manager [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Not allocating networking since 'none' was specified. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 687.563429] env[62183]: INFO nova.compute.manager [-] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Took 1.02 seconds to deallocate network for instance. [ 687.565471] env[62183]: DEBUG nova.compute.claims [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 687.565687] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.849231] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Releasing lock "refresh_cache-14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.849231] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.849231] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.849231] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea56d438-7f27-4872-afe1-ef18eb51d9dc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.858879] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c71df8-4590-4ab6-8676-28e76e908b46 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.883196] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b could not be found. [ 687.883417] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 687.883598] env[62183]: INFO nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 687.883837] env[62183]: DEBUG oslo.service.loopingcall [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.884066] env[62183]: DEBUG nova.compute.manager [-] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.884164] env[62183]: DEBUG nova.network.neutron [-] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.904075] env[62183]: DEBUG nova.network.neutron [-] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.060230] env[62183]: DEBUG nova.compute.manager [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.324062] env[62183]: DEBUG nova.compute.manager [req-443bb769-73d1-4612-80eb-c3040a390c5c req-31b8eaa4-f8c5-40c3-9893-f14f3348db91 service nova] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Received event network-changed-11ed88e3-b166-410c-97a4-6d33a650b04c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.326116] env[62183]: DEBUG nova.compute.manager [req-443bb769-73d1-4612-80eb-c3040a390c5c req-31b8eaa4-f8c5-40c3-9893-f14f3348db91 service nova] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Refreshing instance network info cache due to event network-changed-11ed88e3-b166-410c-97a4-6d33a650b04c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 688.326386] env[62183]: DEBUG oslo_concurrency.lockutils [req-443bb769-73d1-4612-80eb-c3040a390c5c req-31b8eaa4-f8c5-40c3-9893-f14f3348db91 service nova] Acquiring lock "refresh_cache-14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.326538] env[62183]: DEBUG oslo_concurrency.lockutils [req-443bb769-73d1-4612-80eb-c3040a390c5c req-31b8eaa4-f8c5-40c3-9893-f14f3348db91 service nova] Acquired lock "refresh_cache-14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.326698] env[62183]: DEBUG nova.network.neutron [req-443bb769-73d1-4612-80eb-c3040a390c5c req-31b8eaa4-f8c5-40c3-9893-f14f3348db91 service nova] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Refreshing network info cache for port 11ed88e3-b166-410c-97a4-6d33a650b04c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 688.406578] env[62183]: DEBUG nova.network.neutron [-] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.431796] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d3315d-2731-4d43-9d76-0b0fd58a4851 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.440198] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665fc7a2-4dcf-4c6a-8932-7fa459b63cf5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.475981] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284f805a-eb75-4678-8a48-df19d4907672 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.483528] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ebba7d-31f2-472f-b642-4a61d796e08d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.496690] env[62183]: DEBUG nova.compute.provider_tree [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.855620] env[62183]: DEBUG nova.network.neutron [req-443bb769-73d1-4612-80eb-c3040a390c5c req-31b8eaa4-f8c5-40c3-9893-f14f3348db91 service nova] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.909027] env[62183]: INFO nova.compute.manager [-] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Took 1.02 seconds to deallocate network for instance. [ 688.912315] env[62183]: DEBUG nova.compute.claims [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 688.912315] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.993136] env[62183]: DEBUG nova.network.neutron [req-443bb769-73d1-4612-80eb-c3040a390c5c req-31b8eaa4-f8c5-40c3-9893-f14f3348db91 service nova] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.001799] env[62183]: DEBUG nova.scheduler.client.report [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 689.076139] env[62183]: DEBUG nova.compute.manager [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.101530] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.102168] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.102168] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.102168] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.102330] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.102390] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.102606] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.102793] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.103446] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.103446] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.103446] env[62183]: DEBUG nova.virt.hardware [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.104177] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c8eae4-5ddc-4084-af9f-ece6f04820ae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.113219] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9df97c7-09cb-488a-b07c-3dde061f670e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.125793] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 689.131538] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Creating folder: Project (d426e86ee03c46dc938550b5a0e51101). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 689.131780] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a52950f-5d77-4f23-aed3-2d92d7d05a60 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.148133] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Created folder: Project (d426e86ee03c46dc938550b5a0e51101) in parent group-v294392. [ 689.148601] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Creating folder: Instances. Parent ref: group-v294405. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 689.148601] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd85d78d-753b-4fc0-a456-d08fdaa3e26a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.159199] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Created folder: Instances in parent group-v294405. [ 689.159426] env[62183]: DEBUG oslo.service.loopingcall [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.159625] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 689.159920] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e4e72235-4f55-40e7-b99d-26a101f59e8d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.176772] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 689.176772] env[62183]: value = "task-1386911" [ 689.176772] env[62183]: _type = "Task" [ 689.176772] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.184742] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386911, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.496141] env[62183]: DEBUG oslo_concurrency.lockutils [req-443bb769-73d1-4612-80eb-c3040a390c5c req-31b8eaa4-f8c5-40c3-9893-f14f3348db91 service nova] Releasing lock "refresh_cache-14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.496436] env[62183]: DEBUG nova.compute.manager [req-443bb769-73d1-4612-80eb-c3040a390c5c req-31b8eaa4-f8c5-40c3-9893-f14f3348db91 service nova] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Received event network-vif-deleted-11ed88e3-b166-410c-97a4-6d33a650b04c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.507492] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.508017] env[62183]: DEBUG nova.compute.manager [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 689.512028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.127s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.693642] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386911, 'name': CreateVM_Task, 'duration_secs': 0.250152} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.693829] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 689.694279] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.694469] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.694831] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 689.695101] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e585bbd5-ff0b-43d4-bbde-d4e0bb2af1c5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.699742] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 689.699742] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]521fc6f0-83f2-3934-cb53-81c4e0cc6123" [ 689.699742] env[62183]: _type = "Task" [ 689.699742] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.707246] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521fc6f0-83f2-3934-cb53-81c4e0cc6123, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.017689] env[62183]: DEBUG nova.compute.utils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.023612] env[62183]: DEBUG nova.compute.manager [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Not allocating networking since 'none' was specified. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 690.209743] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521fc6f0-83f2-3934-cb53-81c4e0cc6123, 'name': SearchDatastore_Task, 'duration_secs': 0.012799} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.212194] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.212450] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 690.212673] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.212816] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.212991] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 690.213566] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d6a2573-d1ae-4efe-bf42-4da5c6c0cc9b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.222496] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 690.224265] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 690.224265] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8eceb42-f7f0-412c-8033-67220411525e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.235593] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 690.235593] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52750c38-ca76-c75d-4976-8a258e1caded" [ 690.235593] env[62183]: _type = "Task" [ 690.235593] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.251249] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52750c38-ca76-c75d-4976-8a258e1caded, 'name': SearchDatastore_Task, 'duration_secs': 0.010095} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.252133] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3321013-acc4-4f76-b5c6-65e495e582ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.260246] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 690.260246] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]521ec203-9e32-8b5f-4791-61e8bc68bfbc" [ 690.260246] env[62183]: _type = "Task" [ 690.260246] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.268754] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521ec203-9e32-8b5f-4791-61e8bc68bfbc, 'name': SearchDatastore_Task} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.269027] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.269865] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 4af88a37-5aa2-47af-9dd9-8233b1bbf077/4af88a37-5aa2-47af-9dd9-8233b1bbf077.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 690.270152] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd36a8a2-1bb5-4338-9882-dd9829f1b034 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.279738] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 690.279738] env[62183]: value = "task-1386912" [ 690.279738] env[62183]: _type = "Task" [ 690.279738] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.289186] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386912, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.371016] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff5e0be-dda5-4cba-a52a-79db58cfaf46 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.376909] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a331d55-01ea-41a8-8fc9-757e254c52fe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.413711] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d24461a-dd46-40ea-9b94-e0375576058c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.421646] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4e60e9-f886-4eb1-81c4-91acf5c711ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.437242] env[62183]: DEBUG nova.compute.provider_tree [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.524013] env[62183]: DEBUG nova.compute.manager [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 690.789706] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386912, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443547} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.789969] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 4af88a37-5aa2-47af-9dd9-8233b1bbf077/4af88a37-5aa2-47af-9dd9-8233b1bbf077.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 690.790199] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 690.790449] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24936c81-6715-4e6f-a524-9263ada14c8d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.796959] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 690.796959] env[62183]: value = "task-1386913" [ 690.796959] env[62183]: _type = "Task" [ 690.796959] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.805218] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386913, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.944080] env[62183]: DEBUG nova.scheduler.client.report [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 690.954983] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Acquiring lock "90be0169-9414-431b-9aa3-82d89bea4bb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.955225] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Lock "90be0169-9414-431b-9aa3-82d89bea4bb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.306305] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386913, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061925} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.307220] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 691.307727] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a541e81-c962-4afe-8127-b9a5376f6c29 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.326524] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 4af88a37-5aa2-47af-9dd9-8233b1bbf077/4af88a37-5aa2-47af-9dd9-8233b1bbf077.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 691.327098] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d03d79d6-a154-4bf9-9f5d-9f1548e46ada {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.346811] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 691.346811] env[62183]: value = "task-1386914" [ 691.346811] env[62183]: _type = "Task" [ 691.346811] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.354277] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386914, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.449401] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.450079] env[62183]: ERROR nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port efcc7c0d-32ee-4834-84b0-a32619083b33, please check neutron logs for more information. [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Traceback (most recent call last): [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self.driver.spawn(context, instance, image_meta, [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] vm_ref = self.build_virtual_machine(instance, [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.450079] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] for vif in network_info: [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] return self._sync_wrapper(fn, *args, **kwargs) [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self.wait() [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self[:] = self._gt.wait() [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] return self._exit_event.wait() [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] result = hub.switch() [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 691.450534] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] return self.greenlet.switch() [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] result = function(*args, **kwargs) [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] return func(*args, **kwargs) [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] raise e [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] nwinfo = self.network_api.allocate_for_instance( [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] created_port_ids = self._update_ports_for_instance( [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] with excutils.save_and_reraise_exception(): [ 691.450998] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] self.force_reraise() [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] raise self.value [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] updated_port = self._update_port( [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] _ensure_no_port_binding_failure(port) [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] raise exception.PortBindingFailed(port_id=port['id']) [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] nova.exception.PortBindingFailed: Binding failed for port efcc7c0d-32ee-4834-84b0-a32619083b33, please check neutron logs for more information. [ 691.451443] env[62183]: ERROR nova.compute.manager [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] [ 691.451745] env[62183]: DEBUG nova.compute.utils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Binding failed for port efcc7c0d-32ee-4834-84b0-a32619083b33, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 691.452124] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.645s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.454922] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Build of instance 5f4edc62-4320-4aa4-ab91-ee77b757bb7c was re-scheduled: Binding failed for port efcc7c0d-32ee-4834-84b0-a32619083b33, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 691.455376] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 691.455604] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "refresh_cache-5f4edc62-4320-4aa4-ab91-ee77b757bb7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.455750] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquired lock "refresh_cache-5f4edc62-4320-4aa4-ab91-ee77b757bb7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.455958] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 691.533236] env[62183]: DEBUG nova.compute.manager [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 691.561515] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 691.561790] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 691.561915] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.562103] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 691.562251] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.562392] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 691.562594] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 691.562746] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 691.562906] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 691.563077] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 691.563247] env[62183]: DEBUG nova.virt.hardware [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.564452] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9226b78d-ac99-43e9-9b40-ba2cdcf33f0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.571799] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54a2224-f1d8-4554-85da-0ea8d8542049 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.585031] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 691.590468] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Creating folder: Project (0fe59a2526884129a3288f769fe85770). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 691.590762] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-62c73b83-0564-41d7-bc85-568c563af36b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.600611] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Created folder: Project (0fe59a2526884129a3288f769fe85770) in parent group-v294392. [ 691.600784] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Creating folder: Instances. Parent ref: group-v294408. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 691.600999] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de930633-517e-43f1-a87d-73305ab3d116 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.609776] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Created folder: Instances in parent group-v294408. [ 691.610042] env[62183]: DEBUG oslo.service.loopingcall [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.610238] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 691.610430] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34ed2aeb-294e-458e-b748-782c017da5c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.626138] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 691.626138] env[62183]: value = "task-1386917" [ 691.626138] env[62183]: _type = "Task" [ 691.626138] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.633364] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386917, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.858940] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386914, 'name': ReconfigVM_Task, 'duration_secs': 0.262813} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.858940] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 4af88a37-5aa2-47af-9dd9-8233b1bbf077/4af88a37-5aa2-47af-9dd9-8233b1bbf077.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 691.858940] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4870ce2e-512c-4963-a1a0-db942d2b0de4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.865837] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 691.865837] env[62183]: value = "task-1386918" [ 691.865837] env[62183]: _type = "Task" [ 691.865837] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.873745] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386918, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.983156] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.080842] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.137809] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386917, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.254474] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397b49c1-8af6-4dcf-9a7a-3b0800819a07 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.261895] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58344062-5642-4f67-b318-c301c6e77ad7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.291356] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e53b8e-1d1f-4b77-a151-2747399e8a58 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.298113] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c8be59-e3df-42a8-8c55-2791d34969d9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.310473] env[62183]: DEBUG nova.compute.provider_tree [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.375922] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386918, 'name': Rename_Task, 'duration_secs': 0.131575} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.376215] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 692.376448] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9714c16-9734-4de9-bac7-50d038d1a152 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.383457] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 692.383457] env[62183]: value = "task-1386919" [ 692.383457] env[62183]: _type = "Task" [ 692.383457] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.390559] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386919, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.585044] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Releasing lock "refresh_cache-5f4edc62-4320-4aa4-ab91-ee77b757bb7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.585044] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 692.585044] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 692.585044] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 692.598576] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.636960] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386917, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.813683] env[62183]: DEBUG nova.scheduler.client.report [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 692.894725] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386919, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.101050] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.138556] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386917, 'name': CreateVM_Task, 'duration_secs': 1.292424} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.138629] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 693.139021] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.139211] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.139484] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 693.139748] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02c432fa-4839-4b4b-809e-83f0e24b706a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.146241] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 693.146241] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]521d9cb2-aa3e-0423-68f7-ef37fb791fb3" [ 693.146241] env[62183]: _type = "Task" [ 693.146241] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.153213] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521d9cb2-aa3e-0423-68f7-ef37fb791fb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.319028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.866s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.319434] env[62183]: ERROR nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 213e9902-c275-4eb1-9741-97c81d49d65f, please check neutron logs for more information. [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Traceback (most recent call last): [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self.driver.spawn(context, instance, image_meta, [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] vm_ref = self.build_virtual_machine(instance, [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.319434] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] for vif in network_info: [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] return self._sync_wrapper(fn, *args, **kwargs) [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self.wait() [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self[:] = self._gt.wait() [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] return self._exit_event.wait() [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] result = hub.switch() [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 693.319800] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] return self.greenlet.switch() [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] result = function(*args, **kwargs) [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] return func(*args, **kwargs) [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] raise e [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] nwinfo = self.network_api.allocate_for_instance( [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] created_port_ids = self._update_ports_for_instance( [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] with excutils.save_and_reraise_exception(): [ 693.320193] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] self.force_reraise() [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] raise self.value [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] updated_port = self._update_port( [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] _ensure_no_port_binding_failure(port) [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] raise exception.PortBindingFailed(port_id=port['id']) [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] nova.exception.PortBindingFailed: Binding failed for port 213e9902-c275-4eb1-9741-97c81d49d65f, please check neutron logs for more information. [ 693.320555] env[62183]: ERROR nova.compute.manager [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] [ 693.320882] env[62183]: DEBUG nova.compute.utils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Binding failed for port 213e9902-c275-4eb1-9741-97c81d49d65f, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 693.321619] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.055s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.324536] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Build of instance f4b29d0e-488e-4d0c-b2cb-1321c40c87ea was re-scheduled: Binding failed for port 213e9902-c275-4eb1-9741-97c81d49d65f, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 693.325091] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 693.325325] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "refresh_cache-f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.325471] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquired lock "refresh_cache-f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.325629] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.394616] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386919, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.603571] env[62183]: INFO nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 5f4edc62-4320-4aa4-ab91-ee77b757bb7c] Took 1.02 seconds to deallocate network for instance. [ 693.658038] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521d9cb2-aa3e-0423-68f7-ef37fb791fb3, 'name': SearchDatastore_Task, 'duration_secs': 0.016943} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.658038] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.658198] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 693.658344] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.658494] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.658667] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 693.658912] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2e4de43-cc60-4c5f-898e-e24a6f3134a1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.666291] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 693.666458] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 693.667177] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d61080d6-31fc-4dfb-bd05-82c3b4f181d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.672057] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 693.672057] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a82902-1e96-fdb8-4212-680798be1db0" [ 693.672057] env[62183]: _type = "Task" [ 693.672057] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.679565] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a82902-1e96-fdb8-4212-680798be1db0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.847622] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.895799] env[62183]: DEBUG oslo_vmware.api [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386919, 'name': PowerOnVM_Task, 'duration_secs': 1.240395} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.896223] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 693.896563] env[62183]: INFO nova.compute.manager [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Took 4.82 seconds to spawn the instance on the hypervisor. [ 693.896762] env[62183]: DEBUG nova.compute.manager [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 693.899805] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918a0510-0096-4e5f-9e2f-798c702778bf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.956355] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.143811] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51859d45-ba79-440c-9b12-033b781779ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.151738] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f124857-5b2d-4d1b-989a-f05952c58c42 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.183995] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49e5728-e131-4dd0-9075-2334ab87dd80 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.192927] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a82902-1e96-fdb8-4212-680798be1db0, 'name': SearchDatastore_Task, 'duration_secs': 0.007959} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.195475] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0eb0e34b-58af-4bb2-a2b0-7cfd5efbbeab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.198477] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e513fbf8-ee21-48e9-b3b9-2f609723ba31 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.205359] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 694.205359] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a7ccd5-d0bc-1cad-05c6-14842bc67a40" [ 694.205359] env[62183]: _type = "Task" [ 694.205359] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.213159] env[62183]: DEBUG nova.compute.provider_tree [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.221925] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a7ccd5-d0bc-1cad-05c6-14842bc67a40, 'name': SearchDatastore_Task, 'duration_secs': 0.009237} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.222693] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.222944] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 8397279a-7a43-4fb0-bc27-17e17e258527/8397279a-7a43-4fb0-bc27-17e17e258527.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 694.223199] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2220e3a3-23ab-4bf2-8890-303fcbe7735e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.229729] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 694.229729] env[62183]: value = "task-1386920" [ 694.229729] env[62183]: _type = "Task" [ 694.229729] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.237040] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386920, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.421179] env[62183]: INFO nova.compute.manager [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Took 26.39 seconds to build instance. [ 694.460162] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Releasing lock "refresh_cache-f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.460162] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 694.460162] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.460162] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.486329] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.639382] env[62183]: INFO nova.scheduler.client.report [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Deleted allocations for instance 5f4edc62-4320-4aa4-ab91-ee77b757bb7c [ 694.718759] env[62183]: DEBUG nova.scheduler.client.report [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 694.740587] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386920, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.411322} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.740839] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 8397279a-7a43-4fb0-bc27-17e17e258527/8397279a-7a43-4fb0-bc27-17e17e258527.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 694.741062] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 694.741414] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d08143ea-cf98-4a95-856d-3cc7086e1dd4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.747489] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 694.747489] env[62183]: value = "task-1386921" [ 694.747489] env[62183]: _type = "Task" [ 694.747489] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.755426] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386921, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.924249] env[62183]: DEBUG oslo_concurrency.lockutils [None req-50d477ad-669f-404e-99fa-bd81b38f0877 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "4af88a37-5aa2-47af-9dd9-8233b1bbf077" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.027s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.989514] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.147637] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "5f4edc62-4320-4aa4-ab91-ee77b757bb7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.758s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.223773] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.902s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.224478] env[62183]: ERROR nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c798064c-381f-4a76-917c-2c5550c0a248, please check neutron logs for more information. [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Traceback (most recent call last): [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self.driver.spawn(context, instance, image_meta, [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] vm_ref = self.build_virtual_machine(instance, [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.224478] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] for vif in network_info: [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] return self._sync_wrapper(fn, *args, **kwargs) [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self.wait() [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self[:] = self._gt.wait() [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] return self._exit_event.wait() [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] result = hub.switch() [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.224790] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] return self.greenlet.switch() [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] result = function(*args, **kwargs) [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] return func(*args, **kwargs) [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] raise e [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] nwinfo = self.network_api.allocate_for_instance( [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] created_port_ids = self._update_ports_for_instance( [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] with excutils.save_and_reraise_exception(): [ 695.225208] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] self.force_reraise() [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] raise self.value [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] updated_port = self._update_port( [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] _ensure_no_port_binding_failure(port) [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] raise exception.PortBindingFailed(port_id=port['id']) [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] nova.exception.PortBindingFailed: Binding failed for port c798064c-381f-4a76-917c-2c5550c0a248, please check neutron logs for more information. [ 695.225528] env[62183]: ERROR nova.compute.manager [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] [ 695.226210] env[62183]: DEBUG nova.compute.utils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Binding failed for port c798064c-381f-4a76-917c-2c5550c0a248, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 695.226782] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.425s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.230040] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Build of instance 0120298f-dc4c-4c13-ab71-70ccd6428dc9 was re-scheduled: Binding failed for port c798064c-381f-4a76-917c-2c5550c0a248, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 695.230656] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 695.230975] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquiring lock "refresh_cache-0120298f-dc4c-4c13-ab71-70ccd6428dc9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.231236] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Acquired lock "refresh_cache-0120298f-dc4c-4c13-ab71-70ccd6428dc9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.231491] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.258534] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386921, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060277} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.258982] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 695.259758] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adfdcf04-1b6b-46c1-bcd3-6c644d8bb50f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.279522] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 8397279a-7a43-4fb0-bc27-17e17e258527/8397279a-7a43-4fb0-bc27-17e17e258527.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 695.280062] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6e89e29-e824-4d52-beff-21a444585229 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.300528] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 695.300528] env[62183]: value = "task-1386922" [ 695.300528] env[62183]: _type = "Task" [ 695.300528] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.310478] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386922, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.429902] env[62183]: DEBUG nova.compute.manager [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.491659] env[62183]: INFO nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: f4b29d0e-488e-4d0c-b2cb-1321c40c87ea] Took 1.03 seconds to deallocate network for instance. [ 695.650139] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.759853] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.812782] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386922, 'name': ReconfigVM_Task, 'duration_secs': 0.325281} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.815170] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 8397279a-7a43-4fb0-bc27-17e17e258527/8397279a-7a43-4fb0-bc27-17e17e258527.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 695.816494] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d07dddfe-4503-4151-8e50-c59b173928dc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.822819] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 695.822819] env[62183]: value = "task-1386923" [ 695.822819] env[62183]: _type = "Task" [ 695.822819] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.831965] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386923, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.852213] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.947625] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.038070] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1249c7cf-75f6-4333-a036-f67e5df1aa77 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.047497] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4831802f-e177-4130-a12e-a29f8b88a786 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.079525] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ecffa5a-11cb-4dc0-b290-a32f6e61c661 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.087077] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329cf971-68a5-4bec-9668-d8eccf4eb0e5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.100209] env[62183]: DEBUG nova.compute.provider_tree [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.169282] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.333352] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386923, 'name': Rename_Task, 'duration_secs': 0.137473} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.333622] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 696.333866] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b86f3786-0d35-4d70-b0a7-052731edeafd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.341650] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 696.341650] env[62183]: value = "task-1386924" [ 696.341650] env[62183]: _type = "Task" [ 696.341650] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.351255] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386924, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.357542] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Releasing lock "refresh_cache-0120298f-dc4c-4c13-ab71-70ccd6428dc9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.357880] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 696.358446] env[62183]: DEBUG nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.358446] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.375991] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.519178] env[62183]: INFO nova.scheduler.client.report [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Deleted allocations for instance f4b29d0e-488e-4d0c-b2cb-1321c40c87ea [ 696.606458] env[62183]: DEBUG nova.scheduler.client.report [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 696.852020] env[62183]: DEBUG oslo_vmware.api [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386924, 'name': PowerOnVM_Task, 'duration_secs': 0.452315} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.852020] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 696.852020] env[62183]: INFO nova.compute.manager [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Took 5.32 seconds to spawn the instance on the hypervisor. [ 696.852020] env[62183]: DEBUG nova.compute.manager [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 696.852344] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd700ef7-110e-4b98-aad0-ee6a53b94d82 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.878788] env[62183]: DEBUG nova.network.neutron [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.027842] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "f4b29d0e-488e-4d0c-b2cb-1321c40c87ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.602s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.111083] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.112143] env[62183]: ERROR nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 91502b7d-747c-46a9-9590-5442cc87cf14, please check neutron logs for more information. [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Traceback (most recent call last): [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self.driver.spawn(context, instance, image_meta, [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] vm_ref = self.build_virtual_machine(instance, [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.112143] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] for vif in network_info: [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] return self._sync_wrapper(fn, *args, **kwargs) [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self.wait() [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self[:] = self._gt.wait() [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] return self._exit_event.wait() [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] result = hub.switch() [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.112517] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] return self.greenlet.switch() [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] result = function(*args, **kwargs) [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] return func(*args, **kwargs) [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] raise e [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] nwinfo = self.network_api.allocate_for_instance( [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] created_port_ids = self._update_ports_for_instance( [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] with excutils.save_and_reraise_exception(): [ 697.112997] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] self.force_reraise() [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] raise self.value [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] updated_port = self._update_port( [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] _ensure_no_port_binding_failure(port) [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] raise exception.PortBindingFailed(port_id=port['id']) [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] nova.exception.PortBindingFailed: Binding failed for port 91502b7d-747c-46a9-9590-5442cc87cf14, please check neutron logs for more information. [ 697.113387] env[62183]: ERROR nova.compute.manager [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] [ 697.113810] env[62183]: DEBUG nova.compute.utils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Binding failed for port 91502b7d-747c-46a9-9590-5442cc87cf14, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.114186] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.934s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.115667] env[62183]: INFO nova.compute.claims [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.118026] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Build of instance 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4 was re-scheduled: Binding failed for port 91502b7d-747c-46a9-9590-5442cc87cf14, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 697.118481] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 697.118699] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquiring lock "refresh_cache-9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.118841] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Acquired lock "refresh_cache-9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.119069] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.366953] env[62183]: INFO nova.compute.manager [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Took 27.28 seconds to build instance. [ 697.381434] env[62183]: INFO nova.compute.manager [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] [instance: 0120298f-dc4c-4c13-ab71-70ccd6428dc9] Took 1.02 seconds to deallocate network for instance. [ 697.530748] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 697.642903] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.769230] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.868891] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5280a26c-e0ad-411d-97ff-59f2cac2ce53 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "8397279a-7a43-4fb0-bc27-17e17e258527" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.802s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.966033] env[62183]: DEBUG nova.compute.manager [None req-f8aed1a9-a737-4817-a933-3022ceab8f62 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 697.966736] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-112f245b-3fa7-4e1a-946b-c0f7964f2871 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.072695] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.160503] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquiring lock "8397279a-7a43-4fb0-bc27-17e17e258527" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.160503] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "8397279a-7a43-4fb0-bc27-17e17e258527" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.160678] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquiring lock "8397279a-7a43-4fb0-bc27-17e17e258527-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.160770] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "8397279a-7a43-4fb0-bc27-17e17e258527-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.160938] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "8397279a-7a43-4fb0-bc27-17e17e258527-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.163595] env[62183]: INFO nova.compute.manager [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Terminating instance [ 698.165474] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquiring lock "refresh_cache-8397279a-7a43-4fb0-bc27-17e17e258527" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.165669] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquired lock "refresh_cache-8397279a-7a43-4fb0-bc27-17e17e258527" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.165841] env[62183]: DEBUG nova.network.neutron [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 698.272211] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Releasing lock "refresh_cache-9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.272476] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 698.272652] env[62183]: DEBUG nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.272815] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.288082] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.372069] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 698.406928] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31eb0885-e425-4a66-94ea-929180b9c4c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.411614] env[62183]: INFO nova.scheduler.client.report [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Deleted allocations for instance 0120298f-dc4c-4c13-ab71-70ccd6428dc9 [ 698.423296] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d40c96e-8fdf-4454-af5c-31960c1ac21c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.456673] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3fa41c1-fd3e-41f7-919b-655c2679c0b3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.465530] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57066f0d-05c4-44b4-a427-18e748827293 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.480101] env[62183]: INFO nova.compute.manager [None req-f8aed1a9-a737-4817-a933-3022ceab8f62 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] instance snapshotting [ 698.480677] env[62183]: DEBUG nova.objects.instance [None req-f8aed1a9-a737-4817-a933-3022ceab8f62 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lazy-loading 'flavor' on Instance uuid 8397279a-7a43-4fb0-bc27-17e17e258527 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 698.482121] env[62183]: DEBUG nova.compute.provider_tree [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.684074] env[62183]: DEBUG nova.network.neutron [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.752656] env[62183]: DEBUG nova.network.neutron [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.792270] env[62183]: DEBUG nova.network.neutron [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.894978] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.923805] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3df5011-8f91-4c25-8bb3-1797976f9ad0 tempest-ListServersNegativeTestJSON-1332904743 tempest-ListServersNegativeTestJSON-1332904743-project-member] Lock "0120298f-dc4c-4c13-ab71-70ccd6428dc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.451s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.986555] env[62183]: DEBUG nova.scheduler.client.report [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 698.990372] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82157ba1-8f9a-4ef3-b16b-b485d3728cfe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.009553] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2625deb3-e0fa-4994-8b2a-eb07a4bd1ed8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.255720] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Releasing lock "refresh_cache-8397279a-7a43-4fb0-bc27-17e17e258527" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.256271] env[62183]: DEBUG nova.compute.manager [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 699.257027] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 699.257453] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c39af1-79e9-49cd-bd62-20505f31fec3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.265692] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 699.265944] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4f1fc59-fde1-4eb8-9d0c-84c3a663f250 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.272359] env[62183]: DEBUG oslo_vmware.api [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 699.272359] env[62183]: value = "task-1386925" [ 699.272359] env[62183]: _type = "Task" [ 699.272359] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.280386] env[62183]: DEBUG oslo_vmware.api [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.294198] env[62183]: INFO nova.compute.manager [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] [instance: 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4] Took 1.02 seconds to deallocate network for instance. [ 699.426508] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.494343] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.494867] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 699.497423] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.255s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.520112] env[62183]: DEBUG nova.compute.manager [None req-f8aed1a9-a737-4817-a933-3022ceab8f62 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Instance disappeared during snapshot {{(pid=62183) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 699.671247] env[62183]: DEBUG nova.compute.manager [None req-f8aed1a9-a737-4817-a933-3022ceab8f62 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Found 0 images (rotation: 2) {{(pid=62183) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 699.784663] env[62183]: DEBUG oslo_vmware.api [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386925, 'name': PowerOffVM_Task, 'duration_secs': 0.117524} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.785520] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 699.785715] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 699.786578] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e2ee16f-3ee9-48a7-8e12-d94bb90a1e0f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.811926] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 699.811926] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 699.813163] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Deleting the datastore file [datastore1] 8397279a-7a43-4fb0-bc27-17e17e258527 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 699.813163] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95b0356e-8998-41d1-98a5-b4a2633f893b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.818707] env[62183]: DEBUG oslo_vmware.api [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for the task: (returnval){ [ 699.818707] env[62183]: value = "task-1386927" [ 699.818707] env[62183]: _type = "Task" [ 699.818707] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.827504] env[62183]: DEBUG oslo_vmware.api [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.965870] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.001819] env[62183]: DEBUG nova.compute.utils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.008092] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 700.008092] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 700.053593] env[62183]: DEBUG nova.policy [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f1687bbb9774073a355ae00b6d57606', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c8e000092584f88ad21ced0f271a2a4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 700.330991] env[62183]: DEBUG oslo_vmware.api [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Task: {'id': task-1386927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.084642} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.332928] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 700.333068] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 700.333310] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 700.333443] env[62183]: INFO nova.compute.manager [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Took 1.08 seconds to destroy the instance on the hypervisor. [ 700.334017] env[62183]: DEBUG oslo.service.loopingcall [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.334017] env[62183]: DEBUG nova.compute.manager [-] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.334017] env[62183]: DEBUG nova.network.neutron [-] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.338143] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Successfully created port: 6d893774-8b7c-414e-808c-d0a5c4ecce3c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.340215] env[62183]: INFO nova.scheduler.client.report [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Deleted allocations for instance 9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4 [ 700.363927] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c57a66-1ba2-426f-94a5-371225baa65f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.370810] env[62183]: DEBUG nova.network.neutron [-] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.372778] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e684d736-85b7-4660-a912-2cdfef9fa1cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.406017] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cdfcbfe-72e3-4998-8687-8558a6d68765 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.413617] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0579347e-ba40-4360-b067-1983c9b1dacc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.427471] env[62183]: DEBUG nova.compute.provider_tree [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.508309] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 700.850932] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c052deba-58e7-45bb-bef7-105b3ee8ab38 tempest-ServersTestMultiNic-213823804 tempest-ServersTestMultiNic-213823804-project-member] Lock "9f17e4b5-28b4-485d-a0f1-b12d74e2d3c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.560s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.878427] env[62183]: DEBUG nova.network.neutron [-] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.932868] env[62183]: DEBUG nova.scheduler.client.report [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 701.339125] env[62183]: DEBUG nova.compute.manager [req-f4197bd7-5560-4a50-8984-427dab11e5cd req-2a06cc29-11ad-4f0d-9cd7-99f38eda5f56 service nova] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Received event network-changed-6d893774-8b7c-414e-808c-d0a5c4ecce3c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.339125] env[62183]: DEBUG nova.compute.manager [req-f4197bd7-5560-4a50-8984-427dab11e5cd req-2a06cc29-11ad-4f0d-9cd7-99f38eda5f56 service nova] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Refreshing instance network info cache due to event network-changed-6d893774-8b7c-414e-808c-d0a5c4ecce3c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 701.339125] env[62183]: DEBUG oslo_concurrency.lockutils [req-f4197bd7-5560-4a50-8984-427dab11e5cd req-2a06cc29-11ad-4f0d-9cd7-99f38eda5f56 service nova] Acquiring lock "refresh_cache-1ad515a9-38ab-4094-9d71-c845dcb72cb6" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.339125] env[62183]: DEBUG oslo_concurrency.lockutils [req-f4197bd7-5560-4a50-8984-427dab11e5cd req-2a06cc29-11ad-4f0d-9cd7-99f38eda5f56 service nova] Acquired lock "refresh_cache-1ad515a9-38ab-4094-9d71-c845dcb72cb6" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.339302] env[62183]: DEBUG nova.network.neutron [req-f4197bd7-5560-4a50-8984-427dab11e5cd req-2a06cc29-11ad-4f0d-9cd7-99f38eda5f56 service nova] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Refreshing network info cache for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 701.354304] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.380899] env[62183]: INFO nova.compute.manager [-] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Took 1.05 seconds to deallocate network for instance. [ 701.449898] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.451091] env[62183]: ERROR nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5926ff4f-0602-40e3-8704-b0dcb398089d, please check neutron logs for more information. [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Traceback (most recent call last): [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self.driver.spawn(context, instance, image_meta, [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] vm_ref = self.build_virtual_machine(instance, [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.451091] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] for vif in network_info: [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] return self._sync_wrapper(fn, *args, **kwargs) [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self.wait() [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self[:] = self._gt.wait() [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] return self._exit_event.wait() [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] result = hub.switch() [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.451636] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] return self.greenlet.switch() [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] result = function(*args, **kwargs) [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] return func(*args, **kwargs) [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] raise e [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] nwinfo = self.network_api.allocate_for_instance( [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] created_port_ids = self._update_ports_for_instance( [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] with excutils.save_and_reraise_exception(): [ 701.451958] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] self.force_reraise() [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] raise self.value [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] updated_port = self._update_port( [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] _ensure_no_port_binding_failure(port) [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] raise exception.PortBindingFailed(port_id=port['id']) [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] nova.exception.PortBindingFailed: Binding failed for port 5926ff4f-0602-40e3-8704-b0dcb398089d, please check neutron logs for more information. [ 701.452291] env[62183]: ERROR nova.compute.manager [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] [ 701.452787] env[62183]: DEBUG nova.compute.utils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Binding failed for port 5926ff4f-0602-40e3-8704-b0dcb398089d, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.453861] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Build of instance f67ab6cd-410b-4ed3-890b-68f3b0f41ec7 was re-scheduled: Binding failed for port 5926ff4f-0602-40e3-8704-b0dcb398089d, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.454313] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.454527] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Acquiring lock "refresh_cache-f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.454665] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Acquired lock "refresh_cache-f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.454814] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.455859] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.890s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.517734] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 701.543194] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.543473] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.543627] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.543807] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.543950] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.544119] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.544325] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.544476] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.545094] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.545094] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.545094] env[62183]: DEBUG nova.virt.hardware [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.546257] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2228d8-0a4d-4422-be46-42a5c44a3c5a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.549561] env[62183]: ERROR nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c, please check neutron logs for more information. [ 701.549561] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.549561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.549561] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.549561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.549561] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.549561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.549561] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.549561] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.549561] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 701.549561] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.549561] env[62183]: ERROR nova.compute.manager raise self.value [ 701.549561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.549561] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.549561] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.549561] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.550094] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.550094] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.550094] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c, please check neutron logs for more information. [ 701.550094] env[62183]: ERROR nova.compute.manager [ 701.550094] env[62183]: Traceback (most recent call last): [ 701.550094] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.550094] env[62183]: listener.cb(fileno) [ 701.550094] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.550094] env[62183]: result = function(*args, **kwargs) [ 701.550094] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.550094] env[62183]: return func(*args, **kwargs) [ 701.550094] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.550094] env[62183]: raise e [ 701.550094] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.550094] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 701.550094] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.550094] env[62183]: created_port_ids = self._update_ports_for_instance( [ 701.550094] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.550094] env[62183]: with excutils.save_and_reraise_exception(): [ 701.550094] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.550094] env[62183]: self.force_reraise() [ 701.550094] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.550094] env[62183]: raise self.value [ 701.550094] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.550094] env[62183]: updated_port = self._update_port( [ 701.550094] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.550094] env[62183]: _ensure_no_port_binding_failure(port) [ 701.550094] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.550094] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.550947] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c, please check neutron logs for more information. [ 701.550947] env[62183]: Removing descriptor: 16 [ 701.555343] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca25699-ed5b-4492-8711-5c6b50a84094 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.569220] env[62183]: ERROR nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c, please check neutron logs for more information. [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Traceback (most recent call last): [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] yield resources [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self.driver.spawn(context, instance, image_meta, [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] vm_ref = self.build_virtual_machine(instance, [ 701.569220] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] for vif in network_info: [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] return self._sync_wrapper(fn, *args, **kwargs) [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self.wait() [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self[:] = self._gt.wait() [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] return self._exit_event.wait() [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.569695] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] current.throw(*self._exc) [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] result = function(*args, **kwargs) [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] return func(*args, **kwargs) [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] raise e [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] nwinfo = self.network_api.allocate_for_instance( [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] created_port_ids = self._update_ports_for_instance( [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] with excutils.save_and_reraise_exception(): [ 701.570127] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self.force_reraise() [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] raise self.value [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] updated_port = self._update_port( [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] _ensure_no_port_binding_failure(port) [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] raise exception.PortBindingFailed(port_id=port['id']) [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] nova.exception.PortBindingFailed: Binding failed for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c, please check neutron logs for more information. [ 701.570548] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] [ 701.570548] env[62183]: INFO nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Terminating instance [ 701.571581] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Acquiring lock "refresh_cache-1ad515a9-38ab-4094-9d71-c845dcb72cb6" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.877078] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.889160] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.912116] env[62183]: DEBUG nova.network.neutron [req-f4197bd7-5560-4a50-8984-427dab11e5cd req-2a06cc29-11ad-4f0d-9cd7-99f38eda5f56 service nova] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.988311] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.043888] env[62183]: DEBUG nova.network.neutron [req-f4197bd7-5560-4a50-8984-427dab11e5cd req-2a06cc29-11ad-4f0d-9cd7-99f38eda5f56 service nova] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.104373] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.306124] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d69d68-810a-41d4-908c-4ce64526f0f4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.313906] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeec0bd0-2080-4a33-b9cd-7534e20cc7b6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.345350] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bb327d-ac0b-4da4-b627-6989a40bd19f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.352840] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3dd67ef-cfe1-4e1e-9396-e6dc09de94a6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.367341] env[62183]: DEBUG nova.compute.provider_tree [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.546249] env[62183]: DEBUG oslo_concurrency.lockutils [req-f4197bd7-5560-4a50-8984-427dab11e5cd req-2a06cc29-11ad-4f0d-9cd7-99f38eda5f56 service nova] Releasing lock "refresh_cache-1ad515a9-38ab-4094-9d71-c845dcb72cb6" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.546777] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Acquired lock "refresh_cache-1ad515a9-38ab-4094-9d71-c845dcb72cb6" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.546862] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.610396] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Releasing lock "refresh_cache-f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.610646] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.610830] env[62183]: DEBUG nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.610998] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.645993] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.874432] env[62183]: DEBUG nova.scheduler.client.report [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 703.069091] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.153199] env[62183]: DEBUG nova.network.neutron [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.235284] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.384926] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.385937] env[62183]: ERROR nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9128c251-e7f0-4964-8590-e1cae8fe8f96, please check neutron logs for more information. [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Traceback (most recent call last): [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self.driver.spawn(context, instance, image_meta, [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] vm_ref = self.build_virtual_machine(instance, [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.385937] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] for vif in network_info: [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] return self._sync_wrapper(fn, *args, **kwargs) [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self.wait() [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self[:] = self._gt.wait() [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] return self._exit_event.wait() [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] result = hub.switch() [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 703.386266] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] return self.greenlet.switch() [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] result = function(*args, **kwargs) [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] return func(*args, **kwargs) [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] raise e [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] nwinfo = self.network_api.allocate_for_instance( [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] created_port_ids = self._update_ports_for_instance( [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] with excutils.save_and_reraise_exception(): [ 703.386611] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] self.force_reraise() [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] raise self.value [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] updated_port = self._update_port( [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] _ensure_no_port_binding_failure(port) [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] raise exception.PortBindingFailed(port_id=port['id']) [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] nova.exception.PortBindingFailed: Binding failed for port 9128c251-e7f0-4964-8590-e1cae8fe8f96, please check neutron logs for more information. [ 703.386953] env[62183]: ERROR nova.compute.manager [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] [ 703.387267] env[62183]: DEBUG nova.compute.utils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Binding failed for port 9128c251-e7f0-4964-8590-e1cae8fe8f96, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.391503] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.479s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.396074] env[62183]: DEBUG nova.compute.manager [req-08f9227e-6920-454b-8a5f-83fa97558215 req-9a6d21b6-6c2a-4b29-9aa4-a9fa79a89318 service nova] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Received event network-vif-deleted-6d893774-8b7c-414e-808c-d0a5c4ecce3c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.399048] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Build of instance ee99cd66-908e-448f-9cce-ebe78a32214d was re-scheduled: Binding failed for port 9128c251-e7f0-4964-8590-e1cae8fe8f96, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 703.399048] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 703.399048] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-ee99cd66-908e-448f-9cce-ebe78a32214d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.399048] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-ee99cd66-908e-448f-9cce-ebe78a32214d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.399341] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.659376] env[62183]: INFO nova.compute.manager [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] [instance: f67ab6cd-410b-4ed3-890b-68f3b0f41ec7] Took 1.05 seconds to deallocate network for instance. [ 703.744472] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Releasing lock "refresh_cache-1ad515a9-38ab-4094-9d71-c845dcb72cb6" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.744651] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.744849] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.745269] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec9d2a03-e35c-4845-82ef-072bca27c6eb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.755351] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc127ec8-9fd5-4440-bce9-7386916d7101 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.779777] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1ad515a9-38ab-4094-9d71-c845dcb72cb6 could not be found. [ 703.780012] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.780212] env[62183]: INFO nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 703.780452] env[62183]: DEBUG oslo.service.loopingcall [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.780671] env[62183]: DEBUG nova.compute.manager [-] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.780765] env[62183]: DEBUG nova.network.neutron [-] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.796813] env[62183]: DEBUG nova.network.neutron [-] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.922421] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.019458] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.261109] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f2714d-71b5-4a82-b58e-c3224af0ecc7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.272068] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c007aa2b-46e7-463b-ba38-c43a3e835bc4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.304083] env[62183]: DEBUG nova.network.neutron [-] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.309363] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc7ce75-aa09-4b4b-a513-61247f84b619 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.314712] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a588f8a-c94f-4e9d-bc9a-22a3dd14b960 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.334236] env[62183]: DEBUG nova.compute.provider_tree [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.525466] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-ee99cd66-908e-448f-9cce-ebe78a32214d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.525622] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 704.525809] env[62183]: DEBUG nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.525976] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.542077] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.705640] env[62183]: INFO nova.scheduler.client.report [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Deleted allocations for instance f67ab6cd-410b-4ed3-890b-68f3b0f41ec7 [ 704.810511] env[62183]: INFO nova.compute.manager [-] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Took 1.03 seconds to deallocate network for instance. [ 704.813415] env[62183]: DEBUG nova.compute.claims [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.813613] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.837019] env[62183]: DEBUG nova.scheduler.client.report [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 705.047942] env[62183]: DEBUG nova.network.neutron [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.216592] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a7a479c3-b249-44ef-a710-65a9dd42529d tempest-ServerTagsTestJSON-612961387 tempest-ServerTagsTestJSON-612961387-project-member] Lock "f67ab6cd-410b-4ed3-890b-68f3b0f41ec7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.746s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.347554] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.953s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.347554] env[62183]: ERROR nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 11ed88e3-b166-410c-97a4-6d33a650b04c, please check neutron logs for more information. [ 705.347554] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Traceback (most recent call last): [ 705.347554] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.347554] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self.driver.spawn(context, instance, image_meta, [ 705.347554] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.347554] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.347554] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.347554] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] vm_ref = self.build_virtual_machine(instance, [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] for vif in network_info: [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] return self._sync_wrapper(fn, *args, **kwargs) [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self.wait() [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self[:] = self._gt.wait() [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] return self._exit_event.wait() [ 705.348017] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] current.throw(*self._exc) [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] result = function(*args, **kwargs) [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] return func(*args, **kwargs) [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] raise e [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] nwinfo = self.network_api.allocate_for_instance( [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] created_port_ids = self._update_ports_for_instance( [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.348523] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] with excutils.save_and_reraise_exception(): [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] self.force_reraise() [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] raise self.value [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] updated_port = self._update_port( [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] _ensure_no_port_binding_failure(port) [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] raise exception.PortBindingFailed(port_id=port['id']) [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] nova.exception.PortBindingFailed: Binding failed for port 11ed88e3-b166-410c-97a4-6d33a650b04c, please check neutron logs for more information. [ 705.348894] env[62183]: ERROR nova.compute.manager [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] [ 705.350163] env[62183]: DEBUG nova.compute.utils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Binding failed for port 11ed88e3-b166-410c-97a4-6d33a650b04c, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.350163] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.401s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.350394] env[62183]: INFO nova.compute.claims [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.355235] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Build of instance 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b was re-scheduled: Binding failed for port 11ed88e3-b166-410c-97a4-6d33a650b04c, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 705.355235] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 705.355235] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Acquiring lock "refresh_cache-14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.355235] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Acquired lock "refresh_cache-14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.355394] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.552996] env[62183]: INFO nova.compute.manager [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: ee99cd66-908e-448f-9cce-ebe78a32214d] Took 1.03 seconds to deallocate network for instance. [ 705.719169] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 705.877052] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.020524] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.258055] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.523324] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Releasing lock "refresh_cache-14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.523773] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 706.524588] env[62183]: DEBUG nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.524588] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.547892] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.588518] env[62183]: INFO nova.scheduler.client.report [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted allocations for instance ee99cd66-908e-448f-9cce-ebe78a32214d [ 706.711064] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249c53d6-39cf-451e-a277-5c877d5ac7cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.720986] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a50fcc-6676-4192-8c81-2ac563f6e407 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.752331] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c803db8b-0b82-49be-9348-5c1ff5a529ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.759901] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a5b2b3-bf6b-44bd-aadc-4a7c532b5b37 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.773310] env[62183]: DEBUG nova.compute.provider_tree [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.055084] env[62183]: DEBUG nova.network.neutron [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.100315] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e3e9947e-0ed0-472a-9533-ce0bdb594988 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "ee99cd66-908e-448f-9cce-ebe78a32214d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.027s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.276146] env[62183]: DEBUG nova.scheduler.client.report [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 707.560023] env[62183]: INFO nova.compute.manager [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] [instance: 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b] Took 1.03 seconds to deallocate network for instance. [ 707.605160] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 707.783985] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.784696] env[62183]: DEBUG nova.compute.manager [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 707.790482] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.621s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.793794] env[62183]: INFO nova.compute.claims [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.125958] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.240818] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Acquiring lock "fe04dc9a-9fda-41e8-b62e-78dc4c026968" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.240986] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Lock "fe04dc9a-9fda-41e8-b62e-78dc4c026968" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.292494] env[62183]: DEBUG nova.compute.utils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.294941] env[62183]: DEBUG nova.compute.manager [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Not allocating networking since 'none' was specified. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 708.339983] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "bd5c642b-05e2-4c96-8e22-54b02fc93263" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.340266] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "bd5c642b-05e2-4c96-8e22-54b02fc93263" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.598207] env[62183]: INFO nova.scheduler.client.report [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Deleted allocations for instance 14d459f5-7d0a-4de5-92f0-878dcdb1fe3b [ 708.797621] env[62183]: DEBUG nova.compute.manager [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.106830] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3626b0dd-f3fc-4319-9310-b24f7ef7b4bd tempest-ServersTestFqdnHostnames-1015662064 tempest-ServersTestFqdnHostnames-1015662064-project-member] Lock "14d459f5-7d0a-4de5-92f0-878dcdb1fe3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.544s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.119461] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae5c37e-58a9-4b37-96a2-ab0a2cd2bcbe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.128098] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1cf8ba-54c9-4167-a032-303bcc1f9c71 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.159154] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6216d86-e97a-4ed3-a90f-7e9f0839d555 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.166890] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95eec763-c23b-4e3b-805b-904caefb3b6e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.181962] env[62183]: DEBUG nova.compute.provider_tree [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.610019] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 709.685038] env[62183]: DEBUG nova.scheduler.client.report [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 709.807406] env[62183]: DEBUG nova.compute.manager [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 709.832809] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 709.833058] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 709.833216] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.833394] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 709.833538] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.833695] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 709.833920] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 709.834090] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 709.834254] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 709.834409] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 709.834574] env[62183]: DEBUG nova.virt.hardware [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.835732] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64b9564-50e1-4290-801f-14f0828cad8b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.843483] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac328b8-d54b-4d53-b919-b81f057ce0a7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.856456] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 709.862154] env[62183]: DEBUG oslo.service.loopingcall [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.862394] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 709.862595] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-618ce8a0-08d6-4174-84f8-12653b7b01c4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.878563] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 709.878563] env[62183]: value = "task-1386932" [ 709.878563] env[62183]: _type = "Task" [ 709.878563] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.886932] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386932, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.135856] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.191262] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.191915] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.199584] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.127s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.201171] env[62183]: INFO nova.compute.claims [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.388890] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386932, 'name': CreateVM_Task, 'duration_secs': 0.333103} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.389325] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 710.389492] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.389646] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.389958] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 710.390531] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a5031ec-2d1e-4312-91d0-7255e4b14173 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.394997] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 710.394997] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52eb3a57-cd83-b2eb-96fc-9048cde779ba" [ 710.394997] env[62183]: _type = "Task" [ 710.394997] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.402488] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52eb3a57-cd83-b2eb-96fc-9048cde779ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.705851] env[62183]: DEBUG nova.compute.utils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.709856] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.710115] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.767568] env[62183]: DEBUG nova.policy [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '854d795d6dce4a5f9ab111ec8b963579', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8438dd309ab4ae982a612030962d060', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.907896] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52eb3a57-cd83-b2eb-96fc-9048cde779ba, 'name': SearchDatastore_Task, 'duration_secs': 0.008826} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.909638] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.910212] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 710.910212] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.910598] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.910598] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 710.910805] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8c96cd3-e640-488e-ad27-cf41a3bc0ade {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.927189] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 710.928010] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 710.928289] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f887f0a-80fb-4812-a2bc-024a2140d894 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.938203] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 710.938203] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52265919-e6e8-1c12-25e0-e1f2e39d3e74" [ 710.938203] env[62183]: _type = "Task" [ 710.938203] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.948857] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52265919-e6e8-1c12-25e0-e1f2e39d3e74, 'name': SearchDatastore_Task, 'duration_secs': 0.008451} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.948857] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76e78f0e-f64c-4925-ac7d-a1f4c2bc9296 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.957138] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 710.957138] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52be197c-eda2-ba23-7f18-303abef438fb" [ 710.957138] env[62183]: _type = "Task" [ 710.957138] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.967025] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52be197c-eda2-ba23-7f18-303abef438fb, 'name': SearchDatastore_Task} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.967025] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.967025] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9/c2182423-e7fe-4ae2-be94-e88a1e49bab9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 710.967025] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09854bee-fb02-4bb6-a378-2c209c3ec660 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.971671] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 710.971671] env[62183]: value = "task-1386934" [ 710.971671] env[62183]: _type = "Task" [ 710.971671] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.980193] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386934, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.135331] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Successfully created port: f7e52a0b-fbb4-435f-b733-bafeee1a764e {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.212252] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.481912] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386934, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476511} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.484707] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9/c2182423-e7fe-4ae2-be94-e88a1e49bab9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 711.485010] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 711.485565] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8abf64d-a18a-4c6d-a7e8-84bef8bb4333 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.492639] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 711.492639] env[62183]: value = "task-1386935" [ 711.492639] env[62183]: _type = "Task" [ 711.492639] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.500913] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386935, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.602242] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8289f5-4749-4b6a-99c0-2f188eab0403 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.610118] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd536157-c737-44c3-8939-a2240a13bd08 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.640796] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7141cd8-f9c6-4ebc-86de-6955b06e5239 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.648510] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b89636-731b-4cd8-926b-47f87ecba2ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.663341] env[62183]: DEBUG nova.compute.provider_tree [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.008278] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386935, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.220217} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.008681] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 712.009818] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bcdac74-6f5c-4687-8093-6964c644e89f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.039757] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9/c2182423-e7fe-4ae2-be94-e88a1e49bab9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 712.041101] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-517a5b66-7935-4f69-a2b9-fdd2279d8754 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.065149] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 712.065149] env[62183]: value = "task-1386936" [ 712.065149] env[62183]: _type = "Task" [ 712.065149] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.073016] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386936, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.167122] env[62183]: DEBUG nova.scheduler.client.report [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 712.224448] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.254126] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.254381] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.254530] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.254732] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.254927] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.255216] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.255530] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.255705] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.255871] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.256043] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.256264] env[62183]: DEBUG nova.virt.hardware [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.257102] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a94be0-5426-4433-b8b6-8bc1deb56828 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.266399] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6716b2b-b686-464d-b604-20dd17490197 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.376121] env[62183]: DEBUG nova.compute.manager [req-c3f57878-ec88-4b50-b93f-f04f82665afb req-2f1cc78d-13ac-482f-a09b-f7b8728d338c service nova] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Received event network-changed-f7e52a0b-fbb4-435f-b733-bafeee1a764e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.376327] env[62183]: DEBUG nova.compute.manager [req-c3f57878-ec88-4b50-b93f-f04f82665afb req-2f1cc78d-13ac-482f-a09b-f7b8728d338c service nova] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Refreshing instance network info cache due to event network-changed-f7e52a0b-fbb4-435f-b733-bafeee1a764e. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 712.376544] env[62183]: DEBUG oslo_concurrency.lockutils [req-c3f57878-ec88-4b50-b93f-f04f82665afb req-2f1cc78d-13ac-482f-a09b-f7b8728d338c service nova] Acquiring lock "refresh_cache-541531de-d96c-47b5-a980-34dfc3e904aa" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.376686] env[62183]: DEBUG oslo_concurrency.lockutils [req-c3f57878-ec88-4b50-b93f-f04f82665afb req-2f1cc78d-13ac-482f-a09b-f7b8728d338c service nova] Acquired lock "refresh_cache-541531de-d96c-47b5-a980-34dfc3e904aa" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.376842] env[62183]: DEBUG nova.network.neutron [req-c3f57878-ec88-4b50-b93f-f04f82665afb req-2f1cc78d-13ac-482f-a09b-f7b8728d338c service nova] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Refreshing network info cache for port f7e52a0b-fbb4-435f-b733-bafeee1a764e {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 712.553205] env[62183]: ERROR nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f7e52a0b-fbb4-435f-b733-bafeee1a764e, please check neutron logs for more information. [ 712.553205] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 712.553205] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.553205] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 712.553205] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.553205] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 712.553205] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.553205] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 712.553205] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.553205] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 712.553205] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.553205] env[62183]: ERROR nova.compute.manager raise self.value [ 712.553205] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.553205] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 712.553205] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.553205] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 712.553750] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.553750] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 712.553750] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f7e52a0b-fbb4-435f-b733-bafeee1a764e, please check neutron logs for more information. [ 712.553750] env[62183]: ERROR nova.compute.manager [ 712.553750] env[62183]: Traceback (most recent call last): [ 712.553750] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 712.553750] env[62183]: listener.cb(fileno) [ 712.553750] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.553750] env[62183]: result = function(*args, **kwargs) [ 712.553750] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.553750] env[62183]: return func(*args, **kwargs) [ 712.553750] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.553750] env[62183]: raise e [ 712.553750] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.553750] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 712.553750] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.553750] env[62183]: created_port_ids = self._update_ports_for_instance( [ 712.553750] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.553750] env[62183]: with excutils.save_and_reraise_exception(): [ 712.553750] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.553750] env[62183]: self.force_reraise() [ 712.553750] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.553750] env[62183]: raise self.value [ 712.553750] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.553750] env[62183]: updated_port = self._update_port( [ 712.553750] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.553750] env[62183]: _ensure_no_port_binding_failure(port) [ 712.553750] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.553750] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.554653] env[62183]: nova.exception.PortBindingFailed: Binding failed for port f7e52a0b-fbb4-435f-b733-bafeee1a764e, please check neutron logs for more information. [ 712.554653] env[62183]: Removing descriptor: 14 [ 712.554653] env[62183]: ERROR nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f7e52a0b-fbb4-435f-b733-bafeee1a764e, please check neutron logs for more information. [ 712.554653] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Traceback (most recent call last): [ 712.554653] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 712.554653] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] yield resources [ 712.554653] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.554653] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self.driver.spawn(context, instance, image_meta, [ 712.554653] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.554653] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.554653] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.554653] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] vm_ref = self.build_virtual_machine(instance, [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] for vif in network_info: [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] return self._sync_wrapper(fn, *args, **kwargs) [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self.wait() [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self[:] = self._gt.wait() [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] return self._exit_event.wait() [ 712.555049] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] result = hub.switch() [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] return self.greenlet.switch() [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] result = function(*args, **kwargs) [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] return func(*args, **kwargs) [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] raise e [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] nwinfo = self.network_api.allocate_for_instance( [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.555471] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] created_port_ids = self._update_ports_for_instance( [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] with excutils.save_and_reraise_exception(): [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self.force_reraise() [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] raise self.value [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] updated_port = self._update_port( [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] _ensure_no_port_binding_failure(port) [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.555891] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] raise exception.PortBindingFailed(port_id=port['id']) [ 712.556298] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] nova.exception.PortBindingFailed: Binding failed for port f7e52a0b-fbb4-435f-b733-bafeee1a764e, please check neutron logs for more information. [ 712.556298] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] [ 712.556298] env[62183]: INFO nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Terminating instance [ 712.557229] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Acquiring lock "refresh_cache-541531de-d96c-47b5-a980-34dfc3e904aa" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.574719] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386936, 'name': ReconfigVM_Task, 'duration_secs': 0.284399} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.574994] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Reconfigured VM instance instance-00000029 to attach disk [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9/c2182423-e7fe-4ae2-be94-e88a1e49bab9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 712.575617] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc45afc6-bb88-4529-a48c-4d677420e8f1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.582435] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 712.582435] env[62183]: value = "task-1386937" [ 712.582435] env[62183]: _type = "Task" [ 712.582435] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.591553] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386937, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.681040] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.681691] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 712.684438] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.790s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.685844] env[62183]: INFO nova.compute.claims [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.896139] env[62183]: DEBUG nova.network.neutron [req-c3f57878-ec88-4b50-b93f-f04f82665afb req-2f1cc78d-13ac-482f-a09b-f7b8728d338c service nova] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.981425] env[62183]: DEBUG nova.network.neutron [req-c3f57878-ec88-4b50-b93f-f04f82665afb req-2f1cc78d-13ac-482f-a09b-f7b8728d338c service nova] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.094917] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386937, 'name': Rename_Task, 'duration_secs': 0.501437} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.095321] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 713.095769] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0159a509-1008-4d34-b47d-9b45841e0b85 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.102264] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 713.102264] env[62183]: value = "task-1386939" [ 713.102264] env[62183]: _type = "Task" [ 713.102264] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.110190] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.190558] env[62183]: DEBUG nova.compute.utils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.194023] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.194198] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 713.260808] env[62183]: DEBUG nova.policy [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59edae5aed2e4ccbb980bc0973890baf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '594f9c41496146c68921dad4c8d99ec4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.484449] env[62183]: DEBUG oslo_concurrency.lockutils [req-c3f57878-ec88-4b50-b93f-f04f82665afb req-2f1cc78d-13ac-482f-a09b-f7b8728d338c service nova] Releasing lock "refresh_cache-541531de-d96c-47b5-a980-34dfc3e904aa" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.484893] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Acquired lock "refresh_cache-541531de-d96c-47b5-a980-34dfc3e904aa" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.485102] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.612963] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386939, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.695382] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 713.708397] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Successfully created port: d89fc11f-2246-439c-8f09-36d206124b7c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.014748] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.068256] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b7dfdb-0670-4ce6-895d-0fedc21239cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.078232] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb461b16-fc9e-4cb2-9b31-73c2a072c932 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.112611] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddb5a7c-6674-4b9a-9acb-19bf05b6ef3b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.121956] env[62183]: DEBUG oslo_vmware.api [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386939, 'name': PowerOnVM_Task, 'duration_secs': 0.524361} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.124076] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 714.124445] env[62183]: INFO nova.compute.manager [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Took 4.32 seconds to spawn the instance on the hypervisor. [ 714.124672] env[62183]: DEBUG nova.compute.manager [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 714.125447] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3561486-0914-4b24-bba9-6bc6c0faae4c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.128824] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5611a2b-4a81-4d0d-af18-43cb045c0087 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.147962] env[62183]: DEBUG nova.compute.provider_tree [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.191987] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.467777] env[62183]: DEBUG nova.compute.manager [req-6714b951-30a0-4dc0-bf2b-0ad7902516a0 req-b3b0b1c0-9934-4e98-b169-5ceec6a9ac98 service nova] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Received event network-vif-deleted-f7e52a0b-fbb4-435f-b733-bafeee1a764e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.656632] env[62183]: DEBUG nova.scheduler.client.report [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 714.665521] env[62183]: INFO nova.compute.manager [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Took 18.73 seconds to build instance. [ 714.694861] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Releasing lock "refresh_cache-541531de-d96c-47b5-a980-34dfc3e904aa" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.695150] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 714.695350] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 714.696010] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6436b246-6548-449a-94e2-1d68be8ca75b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.705409] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852959fd-bbea-4260-9157-3e1234fcb338 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.718582] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.736638] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 541531de-d96c-47b5-a980-34dfc3e904aa could not be found. [ 714.736941] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 714.737153] env[62183]: INFO nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Took 0.04 seconds to destroy the instance on the hypervisor. [ 714.737407] env[62183]: DEBUG oslo.service.loopingcall [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.740167] env[62183]: DEBUG nova.compute.manager [-] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.740287] env[62183]: DEBUG nova.network.neutron [-] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.751357] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.751357] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.751571] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.751809] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.752131] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.752297] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.752533] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.752702] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.752891] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.753065] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.753389] env[62183]: DEBUG nova.virt.hardware [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.754296] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d4440e-681f-4135-9dd2-ebcba306db7e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.763133] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e96e313-b0c5-479e-b4f4-b17d9118b93e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.771945] env[62183]: DEBUG nova.network.neutron [-] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.978009] env[62183]: ERROR nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d89fc11f-2246-439c-8f09-36d206124b7c, please check neutron logs for more information. [ 714.978009] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 714.978009] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.978009] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 714.978009] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.978009] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 714.978009] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.978009] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 714.978009] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.978009] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 714.978009] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.978009] env[62183]: ERROR nova.compute.manager raise self.value [ 714.978009] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.978009] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 714.978009] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.978009] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 714.978864] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.978864] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 714.978864] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d89fc11f-2246-439c-8f09-36d206124b7c, please check neutron logs for more information. [ 714.978864] env[62183]: ERROR nova.compute.manager [ 714.978864] env[62183]: Traceback (most recent call last): [ 714.978864] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 714.978864] env[62183]: listener.cb(fileno) [ 714.978864] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.978864] env[62183]: result = function(*args, **kwargs) [ 714.978864] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.978864] env[62183]: return func(*args, **kwargs) [ 714.978864] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.978864] env[62183]: raise e [ 714.978864] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.978864] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 714.978864] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.978864] env[62183]: created_port_ids = self._update_ports_for_instance( [ 714.978864] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.978864] env[62183]: with excutils.save_and_reraise_exception(): [ 714.978864] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.978864] env[62183]: self.force_reraise() [ 714.978864] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.978864] env[62183]: raise self.value [ 714.978864] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.978864] env[62183]: updated_port = self._update_port( [ 714.978864] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.978864] env[62183]: _ensure_no_port_binding_failure(port) [ 714.978864] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.978864] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 714.980360] env[62183]: nova.exception.PortBindingFailed: Binding failed for port d89fc11f-2246-439c-8f09-36d206124b7c, please check neutron logs for more information. [ 714.980360] env[62183]: Removing descriptor: 14 [ 714.980360] env[62183]: ERROR nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d89fc11f-2246-439c-8f09-36d206124b7c, please check neutron logs for more information. [ 714.980360] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Traceback (most recent call last): [ 714.980360] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 714.980360] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] yield resources [ 714.980360] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.980360] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self.driver.spawn(context, instance, image_meta, [ 714.980360] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 714.980360] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.980360] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.980360] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] vm_ref = self.build_virtual_machine(instance, [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] for vif in network_info: [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] return self._sync_wrapper(fn, *args, **kwargs) [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self.wait() [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self[:] = self._gt.wait() [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] return self._exit_event.wait() [ 714.980946] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] result = hub.switch() [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] return self.greenlet.switch() [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] result = function(*args, **kwargs) [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] return func(*args, **kwargs) [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] raise e [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] nwinfo = self.network_api.allocate_for_instance( [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.981620] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] created_port_ids = self._update_ports_for_instance( [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] with excutils.save_and_reraise_exception(): [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self.force_reraise() [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] raise self.value [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] updated_port = self._update_port( [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] _ensure_no_port_binding_failure(port) [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.982279] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] raise exception.PortBindingFailed(port_id=port['id']) [ 714.982916] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] nova.exception.PortBindingFailed: Binding failed for port d89fc11f-2246-439c-8f09-36d206124b7c, please check neutron logs for more information. [ 714.982916] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] [ 714.982916] env[62183]: INFO nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Terminating instance [ 714.983418] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-82816947-cb4f-4ad7-ad44-1403db312cf8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.983581] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-82816947-cb4f-4ad7-ad44-1403db312cf8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.983753] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 715.165019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.165019] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.169728] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.204s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.171246] env[62183]: INFO nova.compute.claims [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.178225] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c95e634e-88f8-4299-8485-ba51283676fb tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "c2182423-e7fe-4ae2-be94-e88a1e49bab9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.887s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.277362] env[62183]: DEBUG nova.network.neutron [-] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.305220] env[62183]: INFO nova.compute.manager [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Rebuilding instance [ 715.375049] env[62183]: DEBUG nova.compute.manager [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 715.375910] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55cd15b4-eac3-400c-9b5d-b16317b18c42 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.521576] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.573028] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 715.573028] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 715.670788] env[62183]: DEBUG nova.compute.utils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.672293] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.672519] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 715.681983] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.684231] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.775862] env[62183]: DEBUG nova.policy [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f566632d508745469f18df86de087140', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78eb01244db04f7ba71d95b68f90f615', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.781345] env[62183]: INFO nova.compute.manager [-] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Took 1.04 seconds to deallocate network for instance. [ 715.785853] env[62183]: DEBUG nova.compute.claims [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 715.786066] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.887416] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 715.887940] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-895da692-8776-427c-9e80-ce88f19b7ea7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.896425] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 715.896425] env[62183]: value = "task-1386941" [ 715.896425] env[62183]: _type = "Task" [ 715.896425] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.905018] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.083466] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.083466] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 716.176296] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.193438] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-82816947-cb4f-4ad7-ad44-1403db312cf8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.193935] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 716.194292] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 716.197552] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b32e19af-4ab0-428e-963f-df8f625d938b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.210649] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bddbd6a-765f-4e91-87d7-673d94d25637 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.227120] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.240689] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 82816947-cb4f-4ad7-ad44-1403db312cf8 could not be found. [ 716.240689] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 716.240689] env[62183]: INFO nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 716.240689] env[62183]: DEBUG oslo.service.loopingcall [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.240689] env[62183]: DEBUG nova.compute.manager [-] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 716.240689] env[62183]: DEBUG nova.network.neutron [-] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 716.411594] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386941, 'name': PowerOffVM_Task, 'duration_secs': 0.197831} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.411997] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 716.412174] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 716.413177] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f3bfcf-d8ec-437d-ab65-891187f1f958 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.422754] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 716.422976] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0875110-00ce-4f24-be52-71afcd1cda5c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.431828] env[62183]: DEBUG nova.network.neutron [-] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 716.434430] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Successfully created port: a7849fcb-c23b-4e0d-933d-bed9e9eef0d6 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.525566] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 716.527406] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 716.527406] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Deleting the datastore file [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 716.527406] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3339c5d-ea44-4a7a-86d4-0513d1b0c1db {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.533283] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 716.533283] env[62183]: value = "task-1386943" [ 716.533283] env[62183]: _type = "Task" [ 716.533283] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.541645] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386943, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.544241] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b9dbc3-1740-4961-a492-55356aebab37 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.548214] env[62183]: DEBUG nova.compute.manager [req-3ba70361-b70f-48ed-b501-42796a61050f req-f81f7a99-b572-4ae9-84b5-b40a5c1d95b0 service nova] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Received event network-changed-d89fc11f-2246-439c-8f09-36d206124b7c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.548394] env[62183]: DEBUG nova.compute.manager [req-3ba70361-b70f-48ed-b501-42796a61050f req-f81f7a99-b572-4ae9-84b5-b40a5c1d95b0 service nova] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Refreshing instance network info cache due to event network-changed-d89fc11f-2246-439c-8f09-36d206124b7c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 716.548606] env[62183]: DEBUG oslo_concurrency.lockutils [req-3ba70361-b70f-48ed-b501-42796a61050f req-f81f7a99-b572-4ae9-84b5-b40a5c1d95b0 service nova] Acquiring lock "refresh_cache-82816947-cb4f-4ad7-ad44-1403db312cf8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.548786] env[62183]: DEBUG oslo_concurrency.lockutils [req-3ba70361-b70f-48ed-b501-42796a61050f req-f81f7a99-b572-4ae9-84b5-b40a5c1d95b0 service nova] Acquired lock "refresh_cache-82816947-cb4f-4ad7-ad44-1403db312cf8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.548959] env[62183]: DEBUG nova.network.neutron [req-3ba70361-b70f-48ed-b501-42796a61050f req-f81f7a99-b572-4ae9-84b5-b40a5c1d95b0 service nova] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Refreshing network info cache for port d89fc11f-2246-439c-8f09-36d206124b7c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 716.556781] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03888fd-3b9b-4b5e-81b6-bb35d25ec5ce {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.591955] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Didn't find any instances for network info cache update. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 716.592799] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.593570] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cdcd57c-85f3-4d9c-98c2-b6e8c031e671 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.596386] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.596905] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.597496] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.597665] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.597950] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.597950] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 716.598102] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 716.603128] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef75cdc-d1b6-4ea6-ad04-6fb49f3ad283 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.620137] env[62183]: DEBUG nova.compute.provider_tree [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 716.933982] env[62183]: DEBUG nova.network.neutron [-] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.042693] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386943, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135243} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.042943] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 717.043209] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 717.043398] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 717.068603] env[62183]: DEBUG nova.network.neutron [req-3ba70361-b70f-48ed-b501-42796a61050f req-f81f7a99-b572-4ae9-84b5-b40a5c1d95b0 service nova] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.103158] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.147520] env[62183]: ERROR nova.scheduler.client.report [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [req-e956dc8e-a2c1-499a-bd79-4c8e902b2ad6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e956dc8e-a2c1-499a-bd79-4c8e902b2ad6"}]} [ 717.166652] env[62183]: DEBUG nova.scheduler.client.report [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 717.187334] env[62183]: DEBUG nova.scheduler.client.report [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 717.190241] env[62183]: DEBUG nova.compute.provider_tree [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 717.194610] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.197591] env[62183]: DEBUG nova.network.neutron [req-3ba70361-b70f-48ed-b501-42796a61050f req-f81f7a99-b572-4ae9-84b5-b40a5c1d95b0 service nova] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.219915] env[62183]: DEBUG nova.scheduler.client.report [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 717.230175] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.230684] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.230684] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.230772] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.232018] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.232018] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.232018] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.232018] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.232018] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.232249] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.232249] env[62183]: DEBUG nova.virt.hardware [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.232752] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d992ffa3-7c35-455f-b2ab-fa971e48e33a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.240845] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614851e0-b36d-4880-8314-f400f0d46a91 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.254835] env[62183]: DEBUG nova.scheduler.client.report [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 717.436958] env[62183]: INFO nova.compute.manager [-] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Took 1.20 seconds to deallocate network for instance. [ 717.442178] env[62183]: DEBUG nova.compute.claims [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 717.442178] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.578541] env[62183]: ERROR nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6, please check neutron logs for more information. [ 717.578541] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 717.578541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.578541] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 717.578541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.578541] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 717.578541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.578541] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 717.578541] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.578541] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 717.578541] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.578541] env[62183]: ERROR nova.compute.manager raise self.value [ 717.578541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.578541] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 717.578541] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.578541] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 717.579197] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.579197] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 717.579197] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6, please check neutron logs for more information. [ 717.579197] env[62183]: ERROR nova.compute.manager [ 717.580831] env[62183]: Traceback (most recent call last): [ 717.583019] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 717.583019] env[62183]: listener.cb(fileno) [ 717.583019] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.583019] env[62183]: result = function(*args, **kwargs) [ 717.583019] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.583019] env[62183]: return func(*args, **kwargs) [ 717.583019] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.583019] env[62183]: raise e [ 717.583019] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.583019] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 717.583019] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.583019] env[62183]: created_port_ids = self._update_ports_for_instance( [ 717.583019] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.583019] env[62183]: with excutils.save_and_reraise_exception(): [ 717.583019] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.583019] env[62183]: self.force_reraise() [ 717.583019] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.583019] env[62183]: raise self.value [ 717.583019] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.583019] env[62183]: updated_port = self._update_port( [ 717.583019] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.583019] env[62183]: _ensure_no_port_binding_failure(port) [ 717.583019] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.583019] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 717.583019] env[62183]: nova.exception.PortBindingFailed: Binding failed for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6, please check neutron logs for more information. [ 717.583019] env[62183]: Removing descriptor: 16 [ 717.583869] env[62183]: ERROR nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6, please check neutron logs for more information. [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Traceback (most recent call last): [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] yield resources [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self.driver.spawn(context, instance, image_meta, [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] vm_ref = self.build_virtual_machine(instance, [ 717.583869] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] for vif in network_info: [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] return self._sync_wrapper(fn, *args, **kwargs) [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self.wait() [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self[:] = self._gt.wait() [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] return self._exit_event.wait() [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 717.584234] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] result = hub.switch() [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] return self.greenlet.switch() [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] result = function(*args, **kwargs) [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] return func(*args, **kwargs) [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] raise e [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] nwinfo = self.network_api.allocate_for_instance( [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] created_port_ids = self._update_ports_for_instance( [ 717.584630] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] with excutils.save_and_reraise_exception(): [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self.force_reraise() [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] raise self.value [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] updated_port = self._update_port( [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] _ensure_no_port_binding_failure(port) [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] raise exception.PortBindingFailed(port_id=port['id']) [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] nova.exception.PortBindingFailed: Binding failed for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6, please check neutron logs for more information. [ 717.585085] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] [ 717.585548] env[62183]: INFO nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Terminating instance [ 717.589265] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600a4893-4e07-4757-9a8c-2394a2dd6613 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.593120] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Acquiring lock "refresh_cache-a1cabc70-e6c0-4c9b-8302-6ed8805f2182" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.593770] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Acquired lock "refresh_cache-a1cabc70-e6c0-4c9b-8302-6ed8805f2182" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.594084] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.598471] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a82241e-4673-4c09-a39b-30a73c4ec5be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.636553] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca27f12-d7b6-4dd2-b12f-6c938984b71d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.644267] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ad708c-af65-4738-aa54-b65617fde176 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.658591] env[62183]: DEBUG nova.compute.provider_tree [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.701453] env[62183]: DEBUG oslo_concurrency.lockutils [req-3ba70361-b70f-48ed-b501-42796a61050f req-f81f7a99-b572-4ae9-84b5-b40a5c1d95b0 service nova] Releasing lock "refresh_cache-82816947-cb4f-4ad7-ad44-1403db312cf8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.701776] env[62183]: DEBUG nova.compute.manager [req-3ba70361-b70f-48ed-b501-42796a61050f req-f81f7a99-b572-4ae9-84b5-b40a5c1d95b0 service nova] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Received event network-vif-deleted-d89fc11f-2246-439c-8f09-36d206124b7c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.047202] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "27ba0765-152b-4985-927d-99670818a0b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.047202] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "27ba0765-152b-4985-927d-99670818a0b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.089943] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.089943] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.089943] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.089943] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.090181] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.090218] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.090402] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.090555] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.090721] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.090879] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.091066] env[62183]: DEBUG nova.virt.hardware [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.091986] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74fa8d16-5643-45fb-922d-7bd685c39404 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.099895] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44806598-50ee-4799-9a52-651eda2ff284 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.115427] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.121229] env[62183]: DEBUG oslo.service.loopingcall [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.121229] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 718.121371] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d10f517-0752-4de6-b297-947f9cfb0ff0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.133159] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.139939] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.139939] env[62183]: value = "task-1386945" [ 718.139939] env[62183]: _type = "Task" [ 718.139939] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.148145] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386945, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.162208] env[62183]: DEBUG nova.scheduler.client.report [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 718.190235] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.568389] env[62183]: DEBUG nova.compute.manager [req-cbb86153-8ed9-42b6-9623-70bcbd390426 req-14eca60f-f285-4daf-bba0-66438044af98 service nova] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Received event network-changed-a7849fcb-c23b-4e0d-933d-bed9e9eef0d6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.568582] env[62183]: DEBUG nova.compute.manager [req-cbb86153-8ed9-42b6-9623-70bcbd390426 req-14eca60f-f285-4daf-bba0-66438044af98 service nova] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Refreshing instance network info cache due to event network-changed-a7849fcb-c23b-4e0d-933d-bed9e9eef0d6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.568754] env[62183]: DEBUG oslo_concurrency.lockutils [req-cbb86153-8ed9-42b6-9623-70bcbd390426 req-14eca60f-f285-4daf-bba0-66438044af98 service nova] Acquiring lock "refresh_cache-a1cabc70-e6c0-4c9b-8302-6ed8805f2182" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.649857] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386945, 'name': CreateVM_Task, 'duration_secs': 0.335897} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.650043] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 718.650466] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.650641] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.650968] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 718.651232] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad183c73-8f61-4e36-a5d2-3ae2949f500e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.655858] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 718.655858] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52d7f711-e42d-91e7-e334-4e68ee3d2c30" [ 718.655858] env[62183]: _type = "Task" [ 718.655858] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.663862] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d7f711-e42d-91e7-e334-4e68ee3d2c30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.666634] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.497s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.667123] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 718.669574] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.793s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.670946] env[62183]: INFO nova.compute.claims [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.693666] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Releasing lock "refresh_cache-a1cabc70-e6c0-4c9b-8302-6ed8805f2182" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.694106] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.694302] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.694615] env[62183]: DEBUG oslo_concurrency.lockutils [req-cbb86153-8ed9-42b6-9623-70bcbd390426 req-14eca60f-f285-4daf-bba0-66438044af98 service nova] Acquired lock "refresh_cache-a1cabc70-e6c0-4c9b-8302-6ed8805f2182" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.694777] env[62183]: DEBUG nova.network.neutron [req-cbb86153-8ed9-42b6-9623-70bcbd390426 req-14eca60f-f285-4daf-bba0-66438044af98 service nova] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Refreshing network info cache for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.695909] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0de72f4d-0ffa-4089-83b9-ae3f57af89fc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.706513] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8bd03c-55c6-4f04-9ec9-84651da46f48 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.733301] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a1cabc70-e6c0-4c9b-8302-6ed8805f2182 could not be found. [ 718.733301] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 718.733301] env[62183]: INFO nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Took 0.04 seconds to destroy the instance on the hypervisor. [ 718.733301] env[62183]: DEBUG oslo.service.loopingcall [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.733301] env[62183]: DEBUG nova.compute.manager [-] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.733301] env[62183]: DEBUG nova.network.neutron [-] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.755545] env[62183]: DEBUG nova.network.neutron [-] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.166835] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d7f711-e42d-91e7-e334-4e68ee3d2c30, 'name': SearchDatastore_Task, 'duration_secs': 0.009575} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.167178] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.167411] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.167640] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.167787] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.167992] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.168266] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95654024-d37b-4b87-afcc-e50349613fd6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.175371] env[62183]: DEBUG nova.compute.utils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.178641] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.178811] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 719.182121] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.182308] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 719.183304] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a350beb-bd7a-415a-9b3e-97ec4fdf0332 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.188470] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 719.188470] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52c4ce5a-74e9-4433-b5b7-1130712899c8" [ 719.188470] env[62183]: _type = "Task" [ 719.188470] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.196482] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c4ce5a-74e9-4433-b5b7-1130712899c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.218684] env[62183]: DEBUG nova.network.neutron [req-cbb86153-8ed9-42b6-9623-70bcbd390426 req-14eca60f-f285-4daf-bba0-66438044af98 service nova] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.245922] env[62183]: DEBUG nova.policy [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '20aeadc9abc1467eac21bd9ded19356a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd498301db0a947f588b0f299efc9a9db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.258032] env[62183]: DEBUG nova.network.neutron [-] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.304682] env[62183]: DEBUG nova.network.neutron [req-cbb86153-8ed9-42b6-9623-70bcbd390426 req-14eca60f-f285-4daf-bba0-66438044af98 service nova] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.536286] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Successfully created port: f3e8d80a-3ff2-4a54-9885-5a7339cdc21e {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.683370] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 719.700954] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c4ce5a-74e9-4433-b5b7-1130712899c8, 'name': SearchDatastore_Task, 'duration_secs': 0.00924} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.701917] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96c53680-d407-4097-a96a-82b59bc8fff4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.709539] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 719.709539] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f36010-55c3-d489-71d7-3667d6eac882" [ 719.709539] env[62183]: _type = "Task" [ 719.709539] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.721940] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f36010-55c3-d489-71d7-3667d6eac882, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.761797] env[62183]: INFO nova.compute.manager [-] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Took 1.03 seconds to deallocate network for instance. [ 719.764409] env[62183]: DEBUG nova.compute.claims [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 719.764409] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.810460] env[62183]: DEBUG oslo_concurrency.lockutils [req-cbb86153-8ed9-42b6-9623-70bcbd390426 req-14eca60f-f285-4daf-bba0-66438044af98 service nova] Releasing lock "refresh_cache-a1cabc70-e6c0-4c9b-8302-6ed8805f2182" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.810971] env[62183]: DEBUG nova.compute.manager [req-cbb86153-8ed9-42b6-9623-70bcbd390426 req-14eca60f-f285-4daf-bba0-66438044af98 service nova] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Received event network-vif-deleted-a7849fcb-c23b-4e0d-933d-bed9e9eef0d6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.024591] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2008f9e-5685-4aae-9d68-b1f042ec90f6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.032639] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d542b279-e33a-4a6c-bc4d-32fccc0cd684 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.062238] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73320515-7d97-44cb-af27-597c4453b62b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.069482] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e75676c-dac3-4e1d-b8aa-7e37d32a11c8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.082747] env[62183]: DEBUG nova.compute.provider_tree [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.220568] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f36010-55c3-d489-71d7-3667d6eac882, 'name': SearchDatastore_Task, 'duration_secs': 0.01323} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.221148] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.221202] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9/c2182423-e7fe-4ae2-be94-e88a1e49bab9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 720.221469] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4604fb19-f7cd-4b37-874a-3fbe4ea11d98 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.229060] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 720.229060] env[62183]: value = "task-1386947" [ 720.229060] env[62183]: _type = "Task" [ 720.229060] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.238301] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386947, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.509589] env[62183]: ERROR nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e, please check neutron logs for more information. [ 720.509589] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 720.509589] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.509589] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 720.509589] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.509589] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 720.509589] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.509589] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 720.509589] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.509589] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 720.509589] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.509589] env[62183]: ERROR nova.compute.manager raise self.value [ 720.509589] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.509589] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 720.509589] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.509589] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 720.510174] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.510174] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 720.510174] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e, please check neutron logs for more information. [ 720.510174] env[62183]: ERROR nova.compute.manager [ 720.510174] env[62183]: Traceback (most recent call last): [ 720.510174] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 720.510174] env[62183]: listener.cb(fileno) [ 720.510174] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.510174] env[62183]: result = function(*args, **kwargs) [ 720.510174] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.510174] env[62183]: return func(*args, **kwargs) [ 720.510174] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.510174] env[62183]: raise e [ 720.510174] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.510174] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 720.510174] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.510174] env[62183]: created_port_ids = self._update_ports_for_instance( [ 720.510174] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.510174] env[62183]: with excutils.save_and_reraise_exception(): [ 720.510174] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.510174] env[62183]: self.force_reraise() [ 720.510174] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.510174] env[62183]: raise self.value [ 720.510174] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.510174] env[62183]: updated_port = self._update_port( [ 720.510174] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.510174] env[62183]: _ensure_no_port_binding_failure(port) [ 720.510174] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.510174] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 720.511186] env[62183]: nova.exception.PortBindingFailed: Binding failed for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e, please check neutron logs for more information. [ 720.511186] env[62183]: Removing descriptor: 16 [ 720.586128] env[62183]: DEBUG nova.scheduler.client.report [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 720.595576] env[62183]: DEBUG nova.compute.manager [req-2e5e44ba-cd45-46a4-b2e0-ebc9c53db4f6 req-ceffb781-a4cd-48c8-9fef-a8ae07bd6e12 service nova] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Received event network-changed-f3e8d80a-3ff2-4a54-9885-5a7339cdc21e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.595838] env[62183]: DEBUG nova.compute.manager [req-2e5e44ba-cd45-46a4-b2e0-ebc9c53db4f6 req-ceffb781-a4cd-48c8-9fef-a8ae07bd6e12 service nova] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Refreshing instance network info cache due to event network-changed-f3e8d80a-3ff2-4a54-9885-5a7339cdc21e. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 720.596171] env[62183]: DEBUG oslo_concurrency.lockutils [req-2e5e44ba-cd45-46a4-b2e0-ebc9c53db4f6 req-ceffb781-a4cd-48c8-9fef-a8ae07bd6e12 service nova] Acquiring lock "refresh_cache-3cb4be60-b626-4cc9-91eb-2bd9b778f153" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.596285] env[62183]: DEBUG oslo_concurrency.lockutils [req-2e5e44ba-cd45-46a4-b2e0-ebc9c53db4f6 req-ceffb781-a4cd-48c8-9fef-a8ae07bd6e12 service nova] Acquired lock "refresh_cache-3cb4be60-b626-4cc9-91eb-2bd9b778f153" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.596409] env[62183]: DEBUG nova.network.neutron [req-2e5e44ba-cd45-46a4-b2e0-ebc9c53db4f6 req-ceffb781-a4cd-48c8-9fef-a8ae07bd6e12 service nova] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Refreshing network info cache for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 720.695570] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 720.723893] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.725187] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.725187] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.725187] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.725187] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.725187] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.725559] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.725559] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.725559] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.725559] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.725781] env[62183]: DEBUG nova.virt.hardware [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.726657] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739ce5d5-05fa-4802-88ff-088cfb387750 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.742037] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a861d8e1-f60d-42a7-91d1-a40e975ec348 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.745619] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386947, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.755932] env[62183]: ERROR nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e, please check neutron logs for more information. [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Traceback (most recent call last): [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] yield resources [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self.driver.spawn(context, instance, image_meta, [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] vm_ref = self.build_virtual_machine(instance, [ 720.755932] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] for vif in network_info: [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] return self._sync_wrapper(fn, *args, **kwargs) [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self.wait() [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self[:] = self._gt.wait() [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] return self._exit_event.wait() [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.756384] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] current.throw(*self._exc) [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] result = function(*args, **kwargs) [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] return func(*args, **kwargs) [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] raise e [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] nwinfo = self.network_api.allocate_for_instance( [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] created_port_ids = self._update_ports_for_instance( [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] with excutils.save_and_reraise_exception(): [ 720.756822] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self.force_reraise() [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] raise self.value [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] updated_port = self._update_port( [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] _ensure_no_port_binding_failure(port) [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] raise exception.PortBindingFailed(port_id=port['id']) [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] nova.exception.PortBindingFailed: Binding failed for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e, please check neutron logs for more information. [ 720.757255] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] [ 720.757255] env[62183]: INFO nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Terminating instance [ 720.758487] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "refresh_cache-3cb4be60-b626-4cc9-91eb-2bd9b778f153" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.092468] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.093052] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.095704] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.207s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.095921] env[62183]: DEBUG nova.objects.instance [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lazy-loading 'resources' on Instance uuid 8397279a-7a43-4fb0-bc27-17e17e258527 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 721.113052] env[62183]: DEBUG nova.network.neutron [req-2e5e44ba-cd45-46a4-b2e0-ebc9c53db4f6 req-ceffb781-a4cd-48c8-9fef-a8ae07bd6e12 service nova] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.204676] env[62183]: DEBUG nova.network.neutron [req-2e5e44ba-cd45-46a4-b2e0-ebc9c53db4f6 req-ceffb781-a4cd-48c8-9fef-a8ae07bd6e12 service nova] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.240305] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386947, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.606119] env[62183]: DEBUG nova.compute.utils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.608385] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.608806] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 721.647466] env[62183]: DEBUG nova.policy [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e0eee67c7954c80b4e1ea43b0abbcb5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebcc716f958942b588a6bfde78d2c00d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.707279] env[62183]: DEBUG oslo_concurrency.lockutils [req-2e5e44ba-cd45-46a4-b2e0-ebc9c53db4f6 req-ceffb781-a4cd-48c8-9fef-a8ae07bd6e12 service nova] Releasing lock "refresh_cache-3cb4be60-b626-4cc9-91eb-2bd9b778f153" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.707530] env[62183]: DEBUG nova.compute.manager [req-2e5e44ba-cd45-46a4-b2e0-ebc9c53db4f6 req-ceffb781-a4cd-48c8-9fef-a8ae07bd6e12 service nova] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Received event network-vif-deleted-f3e8d80a-3ff2-4a54-9885-5a7339cdc21e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.707895] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquired lock "refresh_cache-3cb4be60-b626-4cc9-91eb-2bd9b778f153" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.708084] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.744137] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386947, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.447516} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.746923] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9/c2182423-e7fe-4ae2-be94-e88a1e49bab9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 721.747158] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.748269] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04dbe040-572a-40e7-8b7b-b96b781c008a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.756110] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 721.756110] env[62183]: value = "task-1386949" [ 721.756110] env[62183]: _type = "Task" [ 721.756110] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.771251] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.913160] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa90d55-e247-40a5-bb15-28da44c25d6d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.921716] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fb883d-10fa-4f31-8bef-9a82ef4cd607 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.958205] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad8a48e-caeb-4b05-a124-5175b5b682a8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.966665] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Successfully created port: 2d47ffe3-d54f-4d24-a23d-c50c46cf3875 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.968896] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a7ab13-fc28-4ea2-ab1f-4ad2e5790b4a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.984049] env[62183]: DEBUG nova.compute.provider_tree [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.112232] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.232692] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.267513] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065745} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.267985] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.268894] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c89c0e-086e-4020-b317-4dd0123c6e05 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.291076] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9/c2182423-e7fe-4ae2-be94-e88a1e49bab9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.291348] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb4d2021-b964-4317-80eb-523bc05f2d45 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.318658] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 722.318658] env[62183]: value = "task-1386950" [ 722.318658] env[62183]: _type = "Task" [ 722.318658] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.331105] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386950, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.387594] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.487080] env[62183]: DEBUG nova.scheduler.client.report [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 722.789254] env[62183]: DEBUG nova.compute.manager [req-389c237d-6d4f-49ee-8b1c-7c695d8d2d95 req-acf26a89-f3b4-4314-be79-a48baaf31181 service nova] [instance: f25ef564-b649-4557-9847-b5d994079d20] Received event network-changed-2d47ffe3-d54f-4d24-a23d-c50c46cf3875 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.789453] env[62183]: DEBUG nova.compute.manager [req-389c237d-6d4f-49ee-8b1c-7c695d8d2d95 req-acf26a89-f3b4-4314-be79-a48baaf31181 service nova] [instance: f25ef564-b649-4557-9847-b5d994079d20] Refreshing instance network info cache due to event network-changed-2d47ffe3-d54f-4d24-a23d-c50c46cf3875. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.789689] env[62183]: DEBUG oslo_concurrency.lockutils [req-389c237d-6d4f-49ee-8b1c-7c695d8d2d95 req-acf26a89-f3b4-4314-be79-a48baaf31181 service nova] Acquiring lock "refresh_cache-f25ef564-b649-4557-9847-b5d994079d20" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.789826] env[62183]: DEBUG oslo_concurrency.lockutils [req-389c237d-6d4f-49ee-8b1c-7c695d8d2d95 req-acf26a89-f3b4-4314-be79-a48baaf31181 service nova] Acquired lock "refresh_cache-f25ef564-b649-4557-9847-b5d994079d20" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.789981] env[62183]: DEBUG nova.network.neutron [req-389c237d-6d4f-49ee-8b1c-7c695d8d2d95 req-acf26a89-f3b4-4314-be79-a48baaf31181 service nova] [instance: f25ef564-b649-4557-9847-b5d994079d20] Refreshing network info cache for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.828761] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386950, 'name': ReconfigVM_Task, 'duration_secs': 0.285027} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.829190] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Reconfigured VM instance instance-00000029 to attach disk [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9/c2182423-e7fe-4ae2-be94-e88a1e49bab9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.829818] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-acdda0ac-ab1e-45cf-97e9-8a5100da8de4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.837707] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 722.837707] env[62183]: value = "task-1386951" [ 722.837707] env[62183]: _type = "Task" [ 722.837707] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.847668] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386951, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.890271] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Releasing lock "refresh_cache-3cb4be60-b626-4cc9-91eb-2bd9b778f153" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.890750] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.891019] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.891381] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6cbdbf89-5a1d-488c-ad1d-b51c09399001 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.902020] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1cd9ba5-25ca-4116-8389-856f0edba782 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.927625] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3cb4be60-b626-4cc9-91eb-2bd9b778f153 could not be found. [ 722.927862] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.928057] env[62183]: INFO nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Took 0.04 seconds to destroy the instance on the hypervisor. [ 722.928745] env[62183]: DEBUG oslo.service.loopingcall [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.929017] env[62183]: DEBUG nova.compute.manager [-] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.929277] env[62183]: DEBUG nova.network.neutron [-] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.947717] env[62183]: DEBUG nova.network.neutron [-] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.965329] env[62183]: ERROR nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875, please check neutron logs for more information. [ 722.965329] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 722.965329] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.965329] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 722.965329] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.965329] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 722.965329] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.965329] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 722.965329] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.965329] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 722.965329] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.965329] env[62183]: ERROR nova.compute.manager raise self.value [ 722.965329] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.965329] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 722.965329] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.965329] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 722.965855] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.965855] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 722.965855] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875, please check neutron logs for more information. [ 722.965855] env[62183]: ERROR nova.compute.manager [ 722.965855] env[62183]: Traceback (most recent call last): [ 722.965855] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 722.965855] env[62183]: listener.cb(fileno) [ 722.965855] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.965855] env[62183]: result = function(*args, **kwargs) [ 722.965855] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.965855] env[62183]: return func(*args, **kwargs) [ 722.965855] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.965855] env[62183]: raise e [ 722.965855] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.965855] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 722.965855] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.965855] env[62183]: created_port_ids = self._update_ports_for_instance( [ 722.965855] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.965855] env[62183]: with excutils.save_and_reraise_exception(): [ 722.965855] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.965855] env[62183]: self.force_reraise() [ 722.965855] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.965855] env[62183]: raise self.value [ 722.965855] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.965855] env[62183]: updated_port = self._update_port( [ 722.965855] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.965855] env[62183]: _ensure_no_port_binding_failure(port) [ 722.965855] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.965855] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 722.966959] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875, please check neutron logs for more information. [ 722.966959] env[62183]: Removing descriptor: 16 [ 722.993036] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.897s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.994747] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.181s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.014286] env[62183]: INFO nova.scheduler.client.report [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Deleted allocations for instance 8397279a-7a43-4fb0-bc27-17e17e258527 [ 723.121680] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.146587] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.146849] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.147014] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.147207] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.147354] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.147525] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.147775] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.147944] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.148131] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.148336] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.148513] env[62183]: DEBUG nova.virt.hardware [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.149456] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77ea24e-1c43-4727-b507-232f5d43a981 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.159306] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92dd0b2e-ad84-40b0-93cf-c0cf41201d15 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.178503] env[62183]: ERROR nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875, please check neutron logs for more information. [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] Traceback (most recent call last): [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] yield resources [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self.driver.spawn(context, instance, image_meta, [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] vm_ref = self.build_virtual_machine(instance, [ 723.178503] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] for vif in network_info: [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] return self._sync_wrapper(fn, *args, **kwargs) [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self.wait() [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self[:] = self._gt.wait() [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] return self._exit_event.wait() [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 723.179180] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] current.throw(*self._exc) [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] result = function(*args, **kwargs) [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] return func(*args, **kwargs) [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] raise e [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] nwinfo = self.network_api.allocate_for_instance( [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] created_port_ids = self._update_ports_for_instance( [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] with excutils.save_and_reraise_exception(): [ 723.179647] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self.force_reraise() [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] raise self.value [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] updated_port = self._update_port( [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] _ensure_no_port_binding_failure(port) [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] raise exception.PortBindingFailed(port_id=port['id']) [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] nova.exception.PortBindingFailed: Binding failed for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875, please check neutron logs for more information. [ 723.180086] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] [ 723.180086] env[62183]: INFO nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Terminating instance [ 723.181151] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-f25ef564-b649-4557-9847-b5d994079d20" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.309261] env[62183]: DEBUG nova.network.neutron [req-389c237d-6d4f-49ee-8b1c-7c695d8d2d95 req-acf26a89-f3b4-4314-be79-a48baaf31181 service nova] [instance: f25ef564-b649-4557-9847-b5d994079d20] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.348589] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386951, 'name': Rename_Task, 'duration_secs': 0.135478} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.348864] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 723.349180] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d1ee408-e947-4389-b8dd-99879af6fda9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.357149] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 723.357149] env[62183]: value = "task-1386952" [ 723.357149] env[62183]: _type = "Task" [ 723.357149] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.366340] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386952, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.424504] env[62183]: DEBUG nova.network.neutron [req-389c237d-6d4f-49ee-8b1c-7c695d8d2d95 req-acf26a89-f3b4-4314-be79-a48baaf31181 service nova] [instance: f25ef564-b649-4557-9847-b5d994079d20] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.451144] env[62183]: DEBUG nova.network.neutron [-] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.524548] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0e6b6139-bcf4-4438-b265-ee10607733f0 tempest-ServersAaction247Test-450028990 tempest-ServersAaction247Test-450028990-project-member] Lock "8397279a-7a43-4fb0-bc27-17e17e258527" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.364s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.781979] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c7a366-4e8d-40d5-9253-b77c4df8be0b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.790099] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b67f506-c49b-43dd-a7a4-a388b05cd459 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.827026] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252a2132-7725-48cd-9b8a-7c8b8feaab90 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.836724] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d787ac-9a9f-44ec-9f59-0569e4ce2fb6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.855480] env[62183]: DEBUG nova.compute.provider_tree [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.866817] env[62183]: DEBUG oslo_vmware.api [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386952, 'name': PowerOnVM_Task, 'duration_secs': 0.435892} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.867815] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 723.868025] env[62183]: DEBUG nova.compute.manager [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 723.868780] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcde6812-03aa-4dcd-ba40-e46e48fe6ce7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.929719] env[62183]: DEBUG oslo_concurrency.lockutils [req-389c237d-6d4f-49ee-8b1c-7c695d8d2d95 req-acf26a89-f3b4-4314-be79-a48baaf31181 service nova] Releasing lock "refresh_cache-f25ef564-b649-4557-9847-b5d994079d20" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.930617] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-f25ef564-b649-4557-9847-b5d994079d20" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.930796] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.955101] env[62183]: INFO nova.compute.manager [-] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Took 1.03 seconds to deallocate network for instance. [ 723.960268] env[62183]: DEBUG nova.compute.claims [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 723.960400] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.358461] env[62183]: DEBUG nova.scheduler.client.report [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 724.387086] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.451275] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.549248] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Acquiring lock "5a559b13-b1e0-414a-896a-e70648ed93d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.549591] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Lock "5a559b13-b1e0-414a-896a-e70648ed93d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.556420] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.811544] env[62183]: DEBUG nova.compute.manager [req-49dd9cbe-e7f3-4e34-a11e-5f08ee027be0 req-7a2e8fee-c607-445d-8acb-c633f17b89a5 service nova] [instance: f25ef564-b649-4557-9847-b5d994079d20] Received event network-vif-deleted-2d47ffe3-d54f-4d24-a23d-c50c46cf3875 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.863544] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.864292] env[62183]: ERROR nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c, please check neutron logs for more information. [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Traceback (most recent call last): [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self.driver.spawn(context, instance, image_meta, [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] vm_ref = self.build_virtual_machine(instance, [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 724.864292] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] for vif in network_info: [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] return self._sync_wrapper(fn, *args, **kwargs) [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self.wait() [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self[:] = self._gt.wait() [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] return self._exit_event.wait() [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] current.throw(*self._exc) [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.864624] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] result = function(*args, **kwargs) [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] return func(*args, **kwargs) [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] raise e [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] nwinfo = self.network_api.allocate_for_instance( [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] created_port_ids = self._update_ports_for_instance( [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] with excutils.save_and_reraise_exception(): [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] self.force_reraise() [ 724.864986] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.865752] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] raise self.value [ 724.865752] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.865752] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] updated_port = self._update_port( [ 724.865752] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.865752] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] _ensure_no_port_binding_failure(port) [ 724.865752] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.865752] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] raise exception.PortBindingFailed(port_id=port['id']) [ 724.865752] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] nova.exception.PortBindingFailed: Binding failed for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c, please check neutron logs for more information. [ 724.865752] env[62183]: ERROR nova.compute.manager [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] [ 724.865752] env[62183]: DEBUG nova.compute.utils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Binding failed for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 724.866241] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.608s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.867916] env[62183]: INFO nova.compute.claims [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.872195] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Build of instance 1ad515a9-38ab-4094-9d71-c845dcb72cb6 was re-scheduled: Binding failed for port 6d893774-8b7c-414e-808c-d0a5c4ecce3c, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 724.872688] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 724.872917] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Acquiring lock "refresh_cache-1ad515a9-38ab-4094-9d71-c845dcb72cb6" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.873077] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Acquired lock "refresh_cache-1ad515a9-38ab-4094-9d71-c845dcb72cb6" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.873237] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.876771] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "07899178-b53c-4b0e-877d-11774286212f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.876771] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "07899178-b53c-4b0e-877d-11774286212f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.059079] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-f25ef564-b649-4557-9847-b5d994079d20" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.059428] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.059627] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.060491] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a7985000-e0ee-4a0e-9fb4-a535a9a66643 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.070094] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f17508b-8617-42ea-93d9-2a5d786cc746 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.096678] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f25ef564-b649-4557-9847-b5d994079d20 could not be found. [ 725.096930] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.097135] env[62183]: INFO nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Took 0.04 seconds to destroy the instance on the hypervisor. [ 725.097382] env[62183]: DEBUG oslo.service.loopingcall [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.097617] env[62183]: DEBUG nova.compute.manager [-] [instance: f25ef564-b649-4557-9847-b5d994079d20] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.097716] env[62183]: DEBUG nova.network.neutron [-] [instance: f25ef564-b649-4557-9847-b5d994079d20] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.111848] env[62183]: DEBUG nova.network.neutron [-] [instance: f25ef564-b649-4557-9847-b5d994079d20] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.257685] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "c2182423-e7fe-4ae2-be94-e88a1e49bab9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.258042] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "c2182423-e7fe-4ae2-be94-e88a1e49bab9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.258137] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "c2182423-e7fe-4ae2-be94-e88a1e49bab9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.258751] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "c2182423-e7fe-4ae2-be94-e88a1e49bab9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.258934] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "c2182423-e7fe-4ae2-be94-e88a1e49bab9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.261387] env[62183]: INFO nova.compute.manager [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Terminating instance [ 725.263156] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "refresh_cache-c2182423-e7fe-4ae2-be94-e88a1e49bab9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.263310] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired lock "refresh_cache-c2182423-e7fe-4ae2-be94-e88a1e49bab9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.263475] env[62183]: DEBUG nova.network.neutron [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.394261] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.478738] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.614833] env[62183]: DEBUG nova.network.neutron [-] [instance: f25ef564-b649-4557-9847-b5d994079d20] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.782431] env[62183]: DEBUG nova.network.neutron [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.834906] env[62183]: DEBUG nova.network.neutron [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.981017] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Releasing lock "refresh_cache-1ad515a9-38ab-4094-9d71-c845dcb72cb6" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.981169] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 725.981356] env[62183]: DEBUG nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.981531] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.000653] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.117724] env[62183]: INFO nova.compute.manager [-] [instance: f25ef564-b649-4557-9847-b5d994079d20] Took 1.02 seconds to deallocate network for instance. [ 726.119955] env[62183]: DEBUG nova.compute.claims [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 726.120147] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.159231] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5336b65-6bd2-4901-86e4-a807ff3c42ed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.168423] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130dec34-6ac7-4ed5-b74e-26fbe7acb194 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.213877] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c0cca0-08ca-4259-8993-9b0f793e0b73 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.226040] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553620ee-c180-43a4-bc9e-74ab9629c179 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.247138] env[62183]: DEBUG nova.compute.provider_tree [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.337907] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Releasing lock "refresh_cache-c2182423-e7fe-4ae2-be94-e88a1e49bab9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.338534] env[62183]: DEBUG nova.compute.manager [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.338821] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.340249] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149f8464-8727-4d97-928a-23c817faf0f5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.351873] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 726.351873] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa1f7bfa-aca0-43ab-ba92-cca2ebf277cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.360180] env[62183]: DEBUG oslo_vmware.api [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 726.360180] env[62183]: value = "task-1386953" [ 726.360180] env[62183]: _type = "Task" [ 726.360180] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.371353] env[62183]: DEBUG oslo_vmware.api [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.503036] env[62183]: DEBUG nova.network.neutron [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.752971] env[62183]: DEBUG nova.scheduler.client.report [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 726.871670] env[62183]: DEBUG oslo_vmware.api [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.006436] env[62183]: INFO nova.compute.manager [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] [instance: 1ad515a9-38ab-4094-9d71-c845dcb72cb6] Took 1.02 seconds to deallocate network for instance. [ 727.258432] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.258968] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.261678] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.136s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.263106] env[62183]: INFO nova.compute.claims [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.371341] env[62183]: DEBUG oslo_vmware.api [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.767465] env[62183]: DEBUG nova.compute.utils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.771470] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 727.771665] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 727.810923] env[62183]: DEBUG nova.policy [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9e6c985f748845c6aeea4212b36b5177', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c17f99f554c44d09344e54dcb8dd799', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 727.871223] env[62183]: DEBUG oslo_vmware.api [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386953, 'name': PowerOffVM_Task, 'duration_secs': 1.434389} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.871485] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 727.871712] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 727.871970] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1307d59a-3cc2-40c0-bc16-d866c52e5fc7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.901661] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 727.901980] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 727.902184] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Deleting the datastore file [datastore1] c2182423-e7fe-4ae2-be94-e88a1e49bab9 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 727.902459] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c0a617c-3aab-4eeb-ab93-bd1947b8ddf3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.909275] env[62183]: DEBUG oslo_vmware.api [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 727.909275] env[62183]: value = "task-1386955" [ 727.909275] env[62183]: _type = "Task" [ 727.909275] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.917935] env[62183]: DEBUG oslo_vmware.api [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386955, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.035732] env[62183]: INFO nova.scheduler.client.report [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Deleted allocations for instance 1ad515a9-38ab-4094-9d71-c845dcb72cb6 [ 728.123719] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Successfully created port: 5cdeefba-3915-4199-9d8f-2518dc131d2e {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.275179] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 728.420344] env[62183]: DEBUG oslo_vmware.api [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386955, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090054} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.422891] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 728.423128] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 728.423317] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.423490] env[62183]: INFO nova.compute.manager [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Took 2.08 seconds to destroy the instance on the hypervisor. [ 728.423726] env[62183]: DEBUG oslo.service.loopingcall [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.424117] env[62183]: DEBUG nova.compute.manager [-] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.424215] env[62183]: DEBUG nova.network.neutron [-] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.452305] env[62183]: DEBUG nova.network.neutron [-] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.543831] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d1e073d8-61ae-47a0-9a7f-799bc658f4df tempest-InstanceActionsV221TestJSON-1634048685 tempest-InstanceActionsV221TestJSON-1634048685-project-member] Lock "1ad515a9-38ab-4094-9d71-c845dcb72cb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.307s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.615810] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47227468-60bd-4b38-91e3-6deee3e5fd21 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.624119] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235a029c-27a4-41ac-ab14-5fda2397e8d1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.656000] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434378d5-571a-4934-a95e-782cf2501ca4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.664696] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5ca148-e134-4500-9e92-28110c9ea4b4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.679118] env[62183]: DEBUG nova.compute.provider_tree [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.955415] env[62183]: DEBUG nova.network.neutron [-] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.985585] env[62183]: DEBUG nova.compute.manager [req-6fc26c04-5347-43f4-900c-1d56ecdea36e req-100337ea-39f9-4713-af47-673611940f30 service nova] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Received event network-changed-5cdeefba-3915-4199-9d8f-2518dc131d2e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.985877] env[62183]: DEBUG nova.compute.manager [req-6fc26c04-5347-43f4-900c-1d56ecdea36e req-100337ea-39f9-4713-af47-673611940f30 service nova] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Refreshing instance network info cache due to event network-changed-5cdeefba-3915-4199-9d8f-2518dc131d2e. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 728.986117] env[62183]: DEBUG oslo_concurrency.lockutils [req-6fc26c04-5347-43f4-900c-1d56ecdea36e req-100337ea-39f9-4713-af47-673611940f30 service nova] Acquiring lock "refresh_cache-b0761551-b2a3-4a38-a32d-f94a36c47d94" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.986265] env[62183]: DEBUG oslo_concurrency.lockutils [req-6fc26c04-5347-43f4-900c-1d56ecdea36e req-100337ea-39f9-4713-af47-673611940f30 service nova] Acquired lock "refresh_cache-b0761551-b2a3-4a38-a32d-f94a36c47d94" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.986425] env[62183]: DEBUG nova.network.neutron [req-6fc26c04-5347-43f4-900c-1d56ecdea36e req-100337ea-39f9-4713-af47-673611940f30 service nova] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Refreshing network info cache for port 5cdeefba-3915-4199-9d8f-2518dc131d2e {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 729.046873] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.127898] env[62183]: ERROR nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5cdeefba-3915-4199-9d8f-2518dc131d2e, please check neutron logs for more information. [ 729.127898] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 729.127898] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.127898] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 729.127898] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.127898] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 729.127898] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.127898] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 729.127898] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.127898] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 729.127898] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.127898] env[62183]: ERROR nova.compute.manager raise self.value [ 729.127898] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.127898] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 729.127898] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.127898] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 729.128705] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.128705] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 729.128705] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5cdeefba-3915-4199-9d8f-2518dc131d2e, please check neutron logs for more information. [ 729.128705] env[62183]: ERROR nova.compute.manager [ 729.128705] env[62183]: Traceback (most recent call last): [ 729.128705] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 729.128705] env[62183]: listener.cb(fileno) [ 729.128705] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.128705] env[62183]: result = function(*args, **kwargs) [ 729.128705] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.128705] env[62183]: return func(*args, **kwargs) [ 729.128705] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.128705] env[62183]: raise e [ 729.128705] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.128705] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 729.128705] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.128705] env[62183]: created_port_ids = self._update_ports_for_instance( [ 729.128705] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.128705] env[62183]: with excutils.save_and_reraise_exception(): [ 729.128705] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.128705] env[62183]: self.force_reraise() [ 729.128705] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.128705] env[62183]: raise self.value [ 729.128705] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.128705] env[62183]: updated_port = self._update_port( [ 729.128705] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.128705] env[62183]: _ensure_no_port_binding_failure(port) [ 729.128705] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.128705] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 729.130147] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 5cdeefba-3915-4199-9d8f-2518dc131d2e, please check neutron logs for more information. [ 729.130147] env[62183]: Removing descriptor: 16 [ 729.185454] env[62183]: DEBUG nova.scheduler.client.report [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 729.288208] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.318008] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.318286] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.318439] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.318618] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.318763] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.318909] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.319126] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.319284] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.319445] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.319631] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.319811] env[62183]: DEBUG nova.virt.hardware [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.320683] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7682884-d03c-4c00-b04d-d5c5282a41d1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.329298] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e21dbcb-7c63-44c3-9376-95ecf99f28a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.343858] env[62183]: ERROR nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5cdeefba-3915-4199-9d8f-2518dc131d2e, please check neutron logs for more information. [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Traceback (most recent call last): [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] yield resources [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self.driver.spawn(context, instance, image_meta, [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] vm_ref = self.build_virtual_machine(instance, [ 729.343858] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] for vif in network_info: [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] return self._sync_wrapper(fn, *args, **kwargs) [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self.wait() [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self[:] = self._gt.wait() [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] return self._exit_event.wait() [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.344274] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] current.throw(*self._exc) [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] result = function(*args, **kwargs) [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] return func(*args, **kwargs) [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] raise e [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] nwinfo = self.network_api.allocate_for_instance( [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] created_port_ids = self._update_ports_for_instance( [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] with excutils.save_and_reraise_exception(): [ 729.344686] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self.force_reraise() [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] raise self.value [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] updated_port = self._update_port( [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] _ensure_no_port_binding_failure(port) [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] raise exception.PortBindingFailed(port_id=port['id']) [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] nova.exception.PortBindingFailed: Binding failed for port 5cdeefba-3915-4199-9d8f-2518dc131d2e, please check neutron logs for more information. [ 729.345156] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] [ 729.345156] env[62183]: INFO nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Terminating instance [ 729.350407] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Acquiring lock "refresh_cache-b0761551-b2a3-4a38-a32d-f94a36c47d94" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.458091] env[62183]: INFO nova.compute.manager [-] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Took 1.03 seconds to deallocate network for instance. [ 729.505207] env[62183]: DEBUG nova.network.neutron [req-6fc26c04-5347-43f4-900c-1d56ecdea36e req-100337ea-39f9-4713-af47-673611940f30 service nova] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.572083] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.664584] env[62183]: DEBUG nova.network.neutron [req-6fc26c04-5347-43f4-900c-1d56ecdea36e req-100337ea-39f9-4713-af47-673611940f30 service nova] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.694139] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.694139] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.694733] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.559s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.696357] env[62183]: INFO nova.compute.claims [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.965396] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.167015] env[62183]: DEBUG oslo_concurrency.lockutils [req-6fc26c04-5347-43f4-900c-1d56ecdea36e req-100337ea-39f9-4713-af47-673611940f30 service nova] Releasing lock "refresh_cache-b0761551-b2a3-4a38-a32d-f94a36c47d94" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.167600] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Acquired lock "refresh_cache-b0761551-b2a3-4a38-a32d-f94a36c47d94" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.167857] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.196518] env[62183]: DEBUG nova.compute.utils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.197734] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.197948] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 730.260186] env[62183]: DEBUG nova.policy [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd0cbdb393e14d97985619e2587cadea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a09410a8dc3a49e2aecc59b0d062e620', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.604339] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Successfully created port: 8739dad5-b146-432a-b96b-6ef9a12276d6 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.694094] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.700733] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.846300] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.043435] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3854a3a1-c8e6-4d1a-b0c8-bcc97679a546 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.055414] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cca327d-eb8b-470f-a537-38fe0fb21922 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.091698] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7eb8ecf-45ff-4fb3-9ca5-8dd8d118293a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.095162] env[62183]: DEBUG nova.compute.manager [req-db632f63-db17-43c0-8135-e3fe08ce69e3 req-f5fb7e1b-acff-47e2-bb10-020be14825bd service nova] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Received event network-vif-deleted-5cdeefba-3915-4199-9d8f-2518dc131d2e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.102505] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9568d293-f9d7-4b47-a297-00f2fe6e80c9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.118298] env[62183]: DEBUG nova.compute.provider_tree [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.349529] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Releasing lock "refresh_cache-b0761551-b2a3-4a38-a32d-f94a36c47d94" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.350264] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 731.350540] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.350859] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3347225d-2660-4e3e-8a52-5b92f995a269 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.362276] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78759235-fe21-42fa-8e05-f0a7a1653289 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.388699] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b0761551-b2a3-4a38-a32d-f94a36c47d94 could not be found. [ 731.388935] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 731.389131] env[62183]: INFO nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Took 0.04 seconds to destroy the instance on the hypervisor. [ 731.389377] env[62183]: DEBUG oslo.service.loopingcall [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.389660] env[62183]: DEBUG nova.compute.manager [-] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.389767] env[62183]: DEBUG nova.network.neutron [-] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.434602] env[62183]: DEBUG nova.network.neutron [-] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.621851] env[62183]: DEBUG nova.scheduler.client.report [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 731.715086] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 731.743718] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.743964] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.744141] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.744319] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.744462] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.744604] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.744808] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.745029] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.745196] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.746808] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.746808] env[62183]: DEBUG nova.virt.hardware [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.746808] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ecd189d-2879-4b21-999b-731647218dec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.755193] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccef0a24-5583-4901-b9d3-5e95e3eb1198 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.937451] env[62183]: DEBUG nova.network.neutron [-] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.129019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.129019] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 732.131770] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.346s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.378157] env[62183]: ERROR nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8739dad5-b146-432a-b96b-6ef9a12276d6, please check neutron logs for more information. [ 732.378157] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.378157] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.378157] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.378157] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.378157] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.378157] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.378157] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.378157] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.378157] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 732.378157] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.378157] env[62183]: ERROR nova.compute.manager raise self.value [ 732.378157] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.378157] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.378157] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.378157] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.378703] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.378703] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.378703] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8739dad5-b146-432a-b96b-6ef9a12276d6, please check neutron logs for more information. [ 732.378703] env[62183]: ERROR nova.compute.manager [ 732.378703] env[62183]: Traceback (most recent call last): [ 732.378703] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.378703] env[62183]: listener.cb(fileno) [ 732.378703] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.378703] env[62183]: result = function(*args, **kwargs) [ 732.378703] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.378703] env[62183]: return func(*args, **kwargs) [ 732.378703] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.378703] env[62183]: raise e [ 732.378703] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.378703] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 732.378703] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.378703] env[62183]: created_port_ids = self._update_ports_for_instance( [ 732.378703] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.378703] env[62183]: with excutils.save_and_reraise_exception(): [ 732.378703] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.378703] env[62183]: self.force_reraise() [ 732.378703] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.378703] env[62183]: raise self.value [ 732.378703] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.378703] env[62183]: updated_port = self._update_port( [ 732.378703] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.378703] env[62183]: _ensure_no_port_binding_failure(port) [ 732.378703] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.378703] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.379617] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 8739dad5-b146-432a-b96b-6ef9a12276d6, please check neutron logs for more information. [ 732.379617] env[62183]: Removing descriptor: 16 [ 732.379617] env[62183]: ERROR nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8739dad5-b146-432a-b96b-6ef9a12276d6, please check neutron logs for more information. [ 732.379617] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Traceback (most recent call last): [ 732.379617] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 732.379617] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] yield resources [ 732.379617] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.379617] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self.driver.spawn(context, instance, image_meta, [ 732.379617] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.379617] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.379617] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.379617] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] vm_ref = self.build_virtual_machine(instance, [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] for vif in network_info: [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] return self._sync_wrapper(fn, *args, **kwargs) [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self.wait() [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self[:] = self._gt.wait() [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] return self._exit_event.wait() [ 732.379988] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] result = hub.switch() [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] return self.greenlet.switch() [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] result = function(*args, **kwargs) [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] return func(*args, **kwargs) [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] raise e [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] nwinfo = self.network_api.allocate_for_instance( [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.380420] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] created_port_ids = self._update_ports_for_instance( [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] with excutils.save_and_reraise_exception(): [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self.force_reraise() [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] raise self.value [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] updated_port = self._update_port( [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] _ensure_no_port_binding_failure(port) [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.380901] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] raise exception.PortBindingFailed(port_id=port['id']) [ 732.381254] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] nova.exception.PortBindingFailed: Binding failed for port 8739dad5-b146-432a-b96b-6ef9a12276d6, please check neutron logs for more information. [ 732.381254] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] [ 732.381254] env[62183]: INFO nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Terminating instance [ 732.382239] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Acquiring lock "refresh_cache-430dcd52-17ce-4710-b9c8-39c28410e356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.382239] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Acquired lock "refresh_cache-430dcd52-17ce-4710-b9c8-39c28410e356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.383014] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.439641] env[62183]: INFO nova.compute.manager [-] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Took 1.05 seconds to deallocate network for instance. [ 732.441852] env[62183]: DEBUG nova.compute.claims [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 732.442052] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.637026] env[62183]: DEBUG nova.compute.utils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.641376] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 732.641546] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 732.720318] env[62183]: DEBUG nova.policy [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9866c89e8ea439d931fdd347873d5a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aaae79668e804348a617d31704363bf1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 732.912409] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.981799] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9596c519-2347-4a9c-bb90-a0dbedf53929 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.000858] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad8db3c-abe0-4ba7-b694-163dc0c4be26 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.032557] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e567d081-1358-4238-b047-604e6f3f9991 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.040772] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451b148b-958e-412e-9d51-437d912079da {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.054376] env[62183]: DEBUG nova.compute.provider_tree [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.056285] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.141866] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 733.167673] env[62183]: DEBUG nova.compute.manager [req-72af79b0-80d1-4910-ba19-95012a6df3ab req-c26fd798-579b-4d48-8e33-265b856498f4 service nova] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Received event network-changed-8739dad5-b146-432a-b96b-6ef9a12276d6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.167975] env[62183]: DEBUG nova.compute.manager [req-72af79b0-80d1-4910-ba19-95012a6df3ab req-c26fd798-579b-4d48-8e33-265b856498f4 service nova] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Refreshing instance network info cache due to event network-changed-8739dad5-b146-432a-b96b-6ef9a12276d6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 733.168499] env[62183]: DEBUG oslo_concurrency.lockutils [req-72af79b0-80d1-4910-ba19-95012a6df3ab req-c26fd798-579b-4d48-8e33-265b856498f4 service nova] Acquiring lock "refresh_cache-430dcd52-17ce-4710-b9c8-39c28410e356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.288019] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Successfully created port: e4527402-f953-4a68-bd4e-d5fff9c83737 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 733.558414] env[62183]: DEBUG nova.scheduler.client.report [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 733.561869] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Releasing lock "refresh_cache-430dcd52-17ce-4710-b9c8-39c28410e356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.562310] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.562558] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.565574] env[62183]: DEBUG oslo_concurrency.lockutils [req-72af79b0-80d1-4910-ba19-95012a6df3ab req-c26fd798-579b-4d48-8e33-265b856498f4 service nova] Acquired lock "refresh_cache-430dcd52-17ce-4710-b9c8-39c28410e356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.565574] env[62183]: DEBUG nova.network.neutron [req-72af79b0-80d1-4910-ba19-95012a6df3ab req-c26fd798-579b-4d48-8e33-265b856498f4 service nova] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Refreshing network info cache for port 8739dad5-b146-432a-b96b-6ef9a12276d6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.565574] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5e9fddc-6ee3-4a94-a5c4-4ef47f0e662b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.575017] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95906978-79eb-4d59-a84a-eeca3e4b5bfb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.603566] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 430dcd52-17ce-4710-b9c8-39c28410e356 could not be found. [ 733.603865] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.604106] env[62183]: INFO nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Took 0.04 seconds to destroy the instance on the hypervisor. [ 733.605250] env[62183]: DEBUG oslo.service.loopingcall [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.605250] env[62183]: DEBUG nova.compute.manager [-] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.605250] env[62183]: DEBUG nova.network.neutron [-] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.633022] env[62183]: DEBUG nova.network.neutron [-] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.064466] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.932s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.065815] env[62183]: ERROR nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f7e52a0b-fbb4-435f-b733-bafeee1a764e, please check neutron logs for more information. [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Traceback (most recent call last): [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self.driver.spawn(context, instance, image_meta, [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] vm_ref = self.build_virtual_machine(instance, [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.065815] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] for vif in network_info: [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] return self._sync_wrapper(fn, *args, **kwargs) [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self.wait() [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self[:] = self._gt.wait() [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] return self._exit_event.wait() [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] result = hub.switch() [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 734.066242] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] return self.greenlet.switch() [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] result = function(*args, **kwargs) [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] return func(*args, **kwargs) [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] raise e [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] nwinfo = self.network_api.allocate_for_instance( [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] created_port_ids = self._update_ports_for_instance( [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] with excutils.save_and_reraise_exception(): [ 734.066705] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] self.force_reraise() [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] raise self.value [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] updated_port = self._update_port( [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] _ensure_no_port_binding_failure(port) [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] raise exception.PortBindingFailed(port_id=port['id']) [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] nova.exception.PortBindingFailed: Binding failed for port f7e52a0b-fbb4-435f-b733-bafeee1a764e, please check neutron logs for more information. [ 734.067095] env[62183]: ERROR nova.compute.manager [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] [ 734.067417] env[62183]: DEBUG nova.compute.utils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Binding failed for port f7e52a0b-fbb4-435f-b733-bafeee1a764e, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 734.069063] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.841s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.071049] env[62183]: INFO nova.compute.claims [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 734.073878] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Build of instance 541531de-d96c-47b5-a980-34dfc3e904aa was re-scheduled: Binding failed for port f7e52a0b-fbb4-435f-b733-bafeee1a764e, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 734.074381] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 734.074608] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Acquiring lock "refresh_cache-541531de-d96c-47b5-a980-34dfc3e904aa" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.074755] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Acquired lock "refresh_cache-541531de-d96c-47b5-a980-34dfc3e904aa" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.074915] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.105495] env[62183]: DEBUG nova.network.neutron [req-72af79b0-80d1-4910-ba19-95012a6df3ab req-c26fd798-579b-4d48-8e33-265b856498f4 service nova] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.135665] env[62183]: DEBUG nova.network.neutron [-] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.155151] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.182699] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.184442] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.184662] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.185071] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.185372] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.185640] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.185982] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.186562] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.186759] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.186947] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.187139] env[62183]: DEBUG nova.virt.hardware [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.188016] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7548d3-8851-481a-b619-0334011f9eaf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.199940] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d1e0aa-5043-456f-8d10-d30e301df229 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.217672] env[62183]: DEBUG nova.network.neutron [req-72af79b0-80d1-4910-ba19-95012a6df3ab req-c26fd798-579b-4d48-8e33-265b856498f4 service nova] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.602694] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.638329] env[62183]: INFO nova.compute.manager [-] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Took 1.03 seconds to deallocate network for instance. [ 734.640741] env[62183]: DEBUG nova.compute.claims [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 734.640904] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.692809] env[62183]: ERROR nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e4527402-f953-4a68-bd4e-d5fff9c83737, please check neutron logs for more information. [ 734.692809] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 734.692809] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.692809] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 734.692809] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.692809] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 734.692809] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.692809] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 734.692809] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.692809] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 734.692809] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.692809] env[62183]: ERROR nova.compute.manager raise self.value [ 734.692809] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.692809] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 734.692809] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.692809] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 734.693385] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.693385] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 734.693385] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e4527402-f953-4a68-bd4e-d5fff9c83737, please check neutron logs for more information. [ 734.693385] env[62183]: ERROR nova.compute.manager [ 734.693385] env[62183]: Traceback (most recent call last): [ 734.693385] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 734.693385] env[62183]: listener.cb(fileno) [ 734.693385] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.693385] env[62183]: result = function(*args, **kwargs) [ 734.693385] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.693385] env[62183]: return func(*args, **kwargs) [ 734.693385] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.693385] env[62183]: raise e [ 734.693385] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.693385] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 734.693385] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.693385] env[62183]: created_port_ids = self._update_ports_for_instance( [ 734.693385] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.693385] env[62183]: with excutils.save_and_reraise_exception(): [ 734.693385] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.693385] env[62183]: self.force_reraise() [ 734.693385] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.693385] env[62183]: raise self.value [ 734.693385] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.693385] env[62183]: updated_port = self._update_port( [ 734.693385] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.693385] env[62183]: _ensure_no_port_binding_failure(port) [ 734.693385] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.693385] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 734.694275] env[62183]: nova.exception.PortBindingFailed: Binding failed for port e4527402-f953-4a68-bd4e-d5fff9c83737, please check neutron logs for more information. [ 734.694275] env[62183]: Removing descriptor: 16 [ 734.694275] env[62183]: ERROR nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e4527402-f953-4a68-bd4e-d5fff9c83737, please check neutron logs for more information. [ 734.694275] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Traceback (most recent call last): [ 734.694275] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 734.694275] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] yield resources [ 734.694275] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.694275] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self.driver.spawn(context, instance, image_meta, [ 734.694275] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.694275] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.694275] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.694275] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] vm_ref = self.build_virtual_machine(instance, [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] for vif in network_info: [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] return self._sync_wrapper(fn, *args, **kwargs) [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self.wait() [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self[:] = self._gt.wait() [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] return self._exit_event.wait() [ 734.694649] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] result = hub.switch() [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] return self.greenlet.switch() [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] result = function(*args, **kwargs) [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] return func(*args, **kwargs) [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] raise e [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] nwinfo = self.network_api.allocate_for_instance( [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.695070] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] created_port_ids = self._update_ports_for_instance( [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] with excutils.save_and_reraise_exception(): [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self.force_reraise() [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] raise self.value [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] updated_port = self._update_port( [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] _ensure_no_port_binding_failure(port) [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.695463] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] raise exception.PortBindingFailed(port_id=port['id']) [ 734.695847] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] nova.exception.PortBindingFailed: Binding failed for port e4527402-f953-4a68-bd4e-d5fff9c83737, please check neutron logs for more information. [ 734.695847] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] [ 734.695847] env[62183]: INFO nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Terminating instance [ 734.696160] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Acquiring lock "refresh_cache-45adf688-561c-49c0-8395-a16bf4e43740" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.696677] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Acquired lock "refresh_cache-45adf688-561c-49c0-8395-a16bf4e43740" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.696852] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.721603] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.722964] env[62183]: DEBUG oslo_concurrency.lockutils [req-72af79b0-80d1-4910-ba19-95012a6df3ab req-c26fd798-579b-4d48-8e33-265b856498f4 service nova] Releasing lock "refresh_cache-430dcd52-17ce-4710-b9c8-39c28410e356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.723226] env[62183]: DEBUG nova.compute.manager [req-72af79b0-80d1-4910-ba19-95012a6df3ab req-c26fd798-579b-4d48-8e33-265b856498f4 service nova] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Received event network-vif-deleted-8739dad5-b146-432a-b96b-6ef9a12276d6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.099491] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Acquiring lock "0dd90b88-ed92-4331-98c5-c7481ac1ae15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.099491] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Lock "0dd90b88-ed92-4331-98c5-c7481ac1ae15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.210214] env[62183]: DEBUG nova.compute.manager [req-e8767eb9-6298-49cf-b6e5-1953a1ec38a6 req-febc084b-ce1f-423c-8aad-504328c657cc service nova] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Received event network-changed-e4527402-f953-4a68-bd4e-d5fff9c83737 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.210492] env[62183]: DEBUG nova.compute.manager [req-e8767eb9-6298-49cf-b6e5-1953a1ec38a6 req-febc084b-ce1f-423c-8aad-504328c657cc service nova] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Refreshing instance network info cache due to event network-changed-e4527402-f953-4a68-bd4e-d5fff9c83737. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.210803] env[62183]: DEBUG oslo_concurrency.lockutils [req-e8767eb9-6298-49cf-b6e5-1953a1ec38a6 req-febc084b-ce1f-423c-8aad-504328c657cc service nova] Acquiring lock "refresh_cache-45adf688-561c-49c0-8395-a16bf4e43740" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.221866] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.225264] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Releasing lock "refresh_cache-541531de-d96c-47b5-a980-34dfc3e904aa" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.225429] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 735.225652] env[62183]: DEBUG nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.225913] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.247644] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.314112] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.402426] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d34ff5-534c-4701-8ed6-338069f671c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.409923] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b882d0-a8d6-43e7-8aee-eb1e16fcf600 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.440868] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf53648-c99d-4d3b-b6be-f0ed8348207b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.448233] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5508d4-1dcb-4b77-9895-10c95238e583 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.461662] env[62183]: DEBUG nova.compute.provider_tree [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.751421] env[62183]: DEBUG nova.network.neutron [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.816387] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Releasing lock "refresh_cache-45adf688-561c-49c0-8395-a16bf4e43740" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.816787] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 735.816978] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.817296] env[62183]: DEBUG oslo_concurrency.lockutils [req-e8767eb9-6298-49cf-b6e5-1953a1ec38a6 req-febc084b-ce1f-423c-8aad-504328c657cc service nova] Acquired lock "refresh_cache-45adf688-561c-49c0-8395-a16bf4e43740" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.817466] env[62183]: DEBUG nova.network.neutron [req-e8767eb9-6298-49cf-b6e5-1953a1ec38a6 req-febc084b-ce1f-423c-8aad-504328c657cc service nova] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Refreshing network info cache for port e4527402-f953-4a68-bd4e-d5fff9c83737 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 735.818443] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3909f179-3fcb-428c-944a-9ce74b1a3784 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.827582] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821abcf2-9de4-40bb-8c3e-4a6b845a456e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.849114] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 45adf688-561c-49c0-8395-a16bf4e43740 could not be found. [ 735.849317] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.849490] env[62183]: INFO nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Took 0.03 seconds to destroy the instance on the hypervisor. [ 735.849739] env[62183]: DEBUG oslo.service.loopingcall [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.849921] env[62183]: DEBUG nova.compute.manager [-] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.850025] env[62183]: DEBUG nova.network.neutron [-] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.864486] env[62183]: DEBUG nova.network.neutron [-] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.965643] env[62183]: DEBUG nova.scheduler.client.report [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 736.257599] env[62183]: INFO nova.compute.manager [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] [instance: 541531de-d96c-47b5-a980-34dfc3e904aa] Took 1.03 seconds to deallocate network for instance. [ 736.338194] env[62183]: DEBUG nova.network.neutron [req-e8767eb9-6298-49cf-b6e5-1953a1ec38a6 req-febc084b-ce1f-423c-8aad-504328c657cc service nova] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.366728] env[62183]: DEBUG nova.network.neutron [-] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.417543] env[62183]: DEBUG nova.network.neutron [req-e8767eb9-6298-49cf-b6e5-1953a1ec38a6 req-febc084b-ce1f-423c-8aad-504328c657cc service nova] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.469851] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.470430] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 736.472995] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.370s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.473135] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.473288] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 736.473940] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.033s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.478984] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363ba0bf-7efc-4e70-956f-d73ae06d0512 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.485424] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eff394d-49f9-4aec-bc53-82bc2070d70a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.498934] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0af475-9d76-46a6-ac76-6c8b99d7a0f5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.505527] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e263040b-c7e5-40cd-ae7d-e4638076f54f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.534285] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181132MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 736.534448] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.869651] env[62183]: INFO nova.compute.manager [-] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Took 1.02 seconds to deallocate network for instance. [ 736.872037] env[62183]: DEBUG nova.compute.claims [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.872222] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.920019] env[62183]: DEBUG oslo_concurrency.lockutils [req-e8767eb9-6298-49cf-b6e5-1953a1ec38a6 req-febc084b-ce1f-423c-8aad-504328c657cc service nova] Releasing lock "refresh_cache-45adf688-561c-49c0-8395-a16bf4e43740" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.920303] env[62183]: DEBUG nova.compute.manager [req-e8767eb9-6298-49cf-b6e5-1953a1ec38a6 req-febc084b-ce1f-423c-8aad-504328c657cc service nova] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Received event network-vif-deleted-e4527402-f953-4a68-bd4e-d5fff9c83737 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.978594] env[62183]: DEBUG nova.compute.utils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.980150] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.980392] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 737.038334] env[62183]: DEBUG nova.policy [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7deeb5ccaca42e29f0cb145a17e1097', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37d4936e2dc34e9d9ed73455081ee4ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 737.270235] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2784ea73-6c0a-4c7e-a484-d05e58217227 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.278341] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf2b6ea-d106-486a-95d4-2ee49ad7d441 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.309840] env[62183]: INFO nova.scheduler.client.report [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Deleted allocations for instance 541531de-d96c-47b5-a980-34dfc3e904aa [ 737.316370] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66be4bf1-3e14-4dc4-a314-ea6757dc56a6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.325884] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bb837f-1db9-40bf-ab3a-05d573ee33f3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.340042] env[62183]: DEBUG nova.compute.provider_tree [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.341061] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Successfully created port: e18a9c5e-d140-412b-8d31-f0329a8b8375 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.487360] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.823437] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c1c9f17a-5be9-4aa9-891f-c3b33a935cc5 tempest-ServerRescueTestJSONUnderV235-1386419921 tempest-ServerRescueTestJSONUnderV235-1386419921-project-member] Lock "541531de-d96c-47b5-a980-34dfc3e904aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.852s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.844694] env[62183]: DEBUG nova.scheduler.client.report [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 738.118840] env[62183]: DEBUG nova.compute.manager [req-29a87198-f22e-4298-bc13-60e149796809 req-c39cfd45-7c06-4abb-89d4-6ee898b6c4e7 service nova] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Received event network-changed-e18a9c5e-d140-412b-8d31-f0329a8b8375 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.119050] env[62183]: DEBUG nova.compute.manager [req-29a87198-f22e-4298-bc13-60e149796809 req-c39cfd45-7c06-4abb-89d4-6ee898b6c4e7 service nova] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Refreshing instance network info cache due to event network-changed-e18a9c5e-d140-412b-8d31-f0329a8b8375. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.119279] env[62183]: DEBUG oslo_concurrency.lockutils [req-29a87198-f22e-4298-bc13-60e149796809 req-c39cfd45-7c06-4abb-89d4-6ee898b6c4e7 service nova] Acquiring lock "refresh_cache-f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.119498] env[62183]: DEBUG oslo_concurrency.lockutils [req-29a87198-f22e-4298-bc13-60e149796809 req-c39cfd45-7c06-4abb-89d4-6ee898b6c4e7 service nova] Acquired lock "refresh_cache-f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.119581] env[62183]: DEBUG nova.network.neutron [req-29a87198-f22e-4298-bc13-60e149796809 req-c39cfd45-7c06-4abb-89d4-6ee898b6c4e7 service nova] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Refreshing network info cache for port e18a9c5e-d140-412b-8d31-f0329a8b8375 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 738.297929] env[62183]: ERROR nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e18a9c5e-d140-412b-8d31-f0329a8b8375, please check neutron logs for more information. [ 738.297929] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 738.297929] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.297929] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 738.297929] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.297929] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 738.297929] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.297929] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 738.297929] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.297929] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 738.297929] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.297929] env[62183]: ERROR nova.compute.manager raise self.value [ 738.297929] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.297929] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 738.297929] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.297929] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 738.298641] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.298641] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 738.298641] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e18a9c5e-d140-412b-8d31-f0329a8b8375, please check neutron logs for more information. [ 738.298641] env[62183]: ERROR nova.compute.manager [ 738.298641] env[62183]: Traceback (most recent call last): [ 738.298641] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 738.298641] env[62183]: listener.cb(fileno) [ 738.298641] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.298641] env[62183]: result = function(*args, **kwargs) [ 738.298641] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.298641] env[62183]: return func(*args, **kwargs) [ 738.298641] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.298641] env[62183]: raise e [ 738.298641] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.298641] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 738.298641] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.298641] env[62183]: created_port_ids = self._update_ports_for_instance( [ 738.298641] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.298641] env[62183]: with excutils.save_and_reraise_exception(): [ 738.298641] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.298641] env[62183]: self.force_reraise() [ 738.298641] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.298641] env[62183]: raise self.value [ 738.298641] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.298641] env[62183]: updated_port = self._update_port( [ 738.298641] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.298641] env[62183]: _ensure_no_port_binding_failure(port) [ 738.298641] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.298641] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 738.299524] env[62183]: nova.exception.PortBindingFailed: Binding failed for port e18a9c5e-d140-412b-8d31-f0329a8b8375, please check neutron logs for more information. [ 738.299524] env[62183]: Removing descriptor: 14 [ 738.325091] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.351024] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.877s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.351555] env[62183]: ERROR nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d89fc11f-2246-439c-8f09-36d206124b7c, please check neutron logs for more information. [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Traceback (most recent call last): [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self.driver.spawn(context, instance, image_meta, [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] vm_ref = self.build_virtual_machine(instance, [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.351555] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] for vif in network_info: [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] return self._sync_wrapper(fn, *args, **kwargs) [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self.wait() [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self[:] = self._gt.wait() [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] return self._exit_event.wait() [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] result = hub.switch() [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 738.351929] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] return self.greenlet.switch() [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] result = function(*args, **kwargs) [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] return func(*args, **kwargs) [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] raise e [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] nwinfo = self.network_api.allocate_for_instance( [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] created_port_ids = self._update_ports_for_instance( [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] with excutils.save_and_reraise_exception(): [ 738.352319] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] self.force_reraise() [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] raise self.value [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] updated_port = self._update_port( [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] _ensure_no_port_binding_failure(port) [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] raise exception.PortBindingFailed(port_id=port['id']) [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] nova.exception.PortBindingFailed: Binding failed for port d89fc11f-2246-439c-8f09-36d206124b7c, please check neutron logs for more information. [ 738.352696] env[62183]: ERROR nova.compute.manager [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] [ 738.353045] env[62183]: DEBUG nova.compute.utils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Binding failed for port d89fc11f-2246-439c-8f09-36d206124b7c, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.353540] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.589s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.359752] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Build of instance 82816947-cb4f-4ad7-ad44-1403db312cf8 was re-scheduled: Binding failed for port d89fc11f-2246-439c-8f09-36d206124b7c, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 738.359752] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 738.359752] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-82816947-cb4f-4ad7-ad44-1403db312cf8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.359752] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-82816947-cb4f-4ad7-ad44-1403db312cf8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.359944] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.496651] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.522085] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.522386] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.522491] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.522665] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.522818] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.522960] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.523178] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.523334] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.523521] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.523697] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.523870] env[62183]: DEBUG nova.virt.hardware [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.525033] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba76eb85-10c0-4d42-be76-e4a5f28c3bfd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.533139] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719ca435-1bcf-4fe1-ae92-55a7af93cb54 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.550627] env[62183]: ERROR nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e18a9c5e-d140-412b-8d31-f0329a8b8375, please check neutron logs for more information. [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Traceback (most recent call last): [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] yield resources [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self.driver.spawn(context, instance, image_meta, [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] vm_ref = self.build_virtual_machine(instance, [ 738.550627] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] for vif in network_info: [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] return self._sync_wrapper(fn, *args, **kwargs) [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self.wait() [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self[:] = self._gt.wait() [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] return self._exit_event.wait() [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.551067] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] current.throw(*self._exc) [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] result = function(*args, **kwargs) [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] return func(*args, **kwargs) [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] raise e [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] nwinfo = self.network_api.allocate_for_instance( [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] created_port_ids = self._update_ports_for_instance( [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] with excutils.save_and_reraise_exception(): [ 738.551505] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self.force_reraise() [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] raise self.value [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] updated_port = self._update_port( [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] _ensure_no_port_binding_failure(port) [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] raise exception.PortBindingFailed(port_id=port['id']) [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] nova.exception.PortBindingFailed: Binding failed for port e18a9c5e-d140-412b-8d31-f0329a8b8375, please check neutron logs for more information. [ 738.551975] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] [ 738.551975] env[62183]: INFO nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Terminating instance [ 738.553203] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "refresh_cache-f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.640346] env[62183]: DEBUG nova.network.neutron [req-29a87198-f22e-4298-bc13-60e149796809 req-c39cfd45-7c06-4abb-89d4-6ee898b6c4e7 service nova] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.736353] env[62183]: DEBUG nova.network.neutron [req-29a87198-f22e-4298-bc13-60e149796809 req-c39cfd45-7c06-4abb-89d4-6ee898b6c4e7 service nova] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.853516] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.877563] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.968160] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.145561] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c78fb5-ae67-42be-8f5d-563cf1a90d33 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.154013] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bca0773-ed8d-4bd4-ac3b-87106ce36ef0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.187828] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f6378c-bfa1-414b-bfc0-e4c0b23119d5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.196173] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da13bb98-ef8a-43fe-83db-c4a689fdf34c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.212678] env[62183]: DEBUG nova.compute.provider_tree [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.238499] env[62183]: DEBUG oslo_concurrency.lockutils [req-29a87198-f22e-4298-bc13-60e149796809 req-c39cfd45-7c06-4abb-89d4-6ee898b6c4e7 service nova] Releasing lock "refresh_cache-f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.238963] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "refresh_cache-f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.239172] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.473168] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-82816947-cb4f-4ad7-ad44-1403db312cf8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.473535] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 739.473696] env[62183]: DEBUG nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.473771] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.490813] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.716690] env[62183]: DEBUG nova.scheduler.client.report [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 739.761533] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.835792] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.993417] env[62183]: DEBUG nova.network.neutron [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.144300] env[62183]: DEBUG nova.compute.manager [req-56a9a655-1987-4631-8c05-aa6f9a47c08c req-096d8087-ce49-4921-bba7-cdbb7fa21148 service nova] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Received event network-vif-deleted-e18a9c5e-d140-412b-8d31-f0329a8b8375 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.224643] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.871s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.225333] env[62183]: ERROR nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6, please check neutron logs for more information. [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Traceback (most recent call last): [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self.driver.spawn(context, instance, image_meta, [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] vm_ref = self.build_virtual_machine(instance, [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.225333] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] for vif in network_info: [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] return self._sync_wrapper(fn, *args, **kwargs) [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self.wait() [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self[:] = self._gt.wait() [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] return self._exit_event.wait() [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] result = hub.switch() [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 740.225685] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] return self.greenlet.switch() [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] result = function(*args, **kwargs) [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] return func(*args, **kwargs) [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] raise e [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] nwinfo = self.network_api.allocate_for_instance( [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] created_port_ids = self._update_ports_for_instance( [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] with excutils.save_and_reraise_exception(): [ 740.226063] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] self.force_reraise() [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] raise self.value [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] updated_port = self._update_port( [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] _ensure_no_port_binding_failure(port) [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] raise exception.PortBindingFailed(port_id=port['id']) [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] nova.exception.PortBindingFailed: Binding failed for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6, please check neutron logs for more information. [ 740.226418] env[62183]: ERROR nova.compute.manager [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] [ 740.226717] env[62183]: DEBUG nova.compute.utils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Binding failed for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.227840] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Build of instance a1cabc70-e6c0-4c9b-8302-6ed8805f2182 was re-scheduled: Binding failed for port a7849fcb-c23b-4e0d-933d-bed9e9eef0d6, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 740.231037] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 740.231037] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Acquiring lock "refresh_cache-a1cabc70-e6c0-4c9b-8302-6ed8805f2182" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.231037] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Acquired lock "refresh_cache-a1cabc70-e6c0-4c9b-8302-6ed8805f2182" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.231037] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.231257] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.269s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.338443] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "refresh_cache-f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.338881] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 740.339082] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 740.339366] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-46947e57-eb46-41a0-a8e7-6d9692f0c04a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.352619] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07d2349-e963-41e7-aee7-e64fabfbcac2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.377533] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f6e8b9ae-a3ea-4164-9a04-a7bb240303b8 could not be found. [ 740.377712] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.377933] env[62183]: INFO nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 740.378256] env[62183]: DEBUG oslo.service.loopingcall [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.378593] env[62183]: DEBUG nova.compute.manager [-] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.378783] env[62183]: DEBUG nova.network.neutron [-] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.394533] env[62183]: DEBUG nova.network.neutron [-] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.497038] env[62183]: INFO nova.compute.manager [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 82816947-cb4f-4ad7-ad44-1403db312cf8] Took 1.02 seconds to deallocate network for instance. [ 740.753773] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.835733] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.897402] env[62183]: DEBUG nova.network.neutron [-] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.996856] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2eed4e-2ba6-4da5-8bb1-45fcd1e88f8d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.011358] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf48865a-b2dd-46f4-b118-f7897a5547ed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.041078] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa9fb99-f0a3-4dd3-bf6f-d030554279b3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.049015] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18da8ced-b440-439e-88cc-717d2fe3952e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.062320] env[62183]: DEBUG nova.compute.provider_tree [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.339315] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Releasing lock "refresh_cache-a1cabc70-e6c0-4c9b-8302-6ed8805f2182" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.339568] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 741.339735] env[62183]: DEBUG nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.339903] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.359591] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.401166] env[62183]: INFO nova.compute.manager [-] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Took 1.02 seconds to deallocate network for instance. [ 741.403878] env[62183]: DEBUG nova.compute.claims [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 741.404075] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.526317] env[62183]: INFO nova.scheduler.client.report [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted allocations for instance 82816947-cb4f-4ad7-ad44-1403db312cf8 [ 741.564982] env[62183]: DEBUG nova.scheduler.client.report [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 741.861922] env[62183]: DEBUG nova.network.neutron [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.034347] env[62183]: DEBUG oslo_concurrency.lockutils [None req-448f707a-94d4-420c-9279-07876d669e4f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "82816947-cb4f-4ad7-ad44-1403db312cf8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.043s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.070571] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.839s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.070571] env[62183]: ERROR nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e, please check neutron logs for more information. [ 742.070571] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Traceback (most recent call last): [ 742.070571] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.070571] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self.driver.spawn(context, instance, image_meta, [ 742.070571] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 742.070571] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.070571] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.070571] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] vm_ref = self.build_virtual_machine(instance, [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] for vif in network_info: [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] return self._sync_wrapper(fn, *args, **kwargs) [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self.wait() [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self[:] = self._gt.wait() [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] return self._exit_event.wait() [ 742.070924] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] current.throw(*self._exc) [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] result = function(*args, **kwargs) [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] return func(*args, **kwargs) [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] raise e [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] nwinfo = self.network_api.allocate_for_instance( [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] created_port_ids = self._update_ports_for_instance( [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.071465] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] with excutils.save_and_reraise_exception(): [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] self.force_reraise() [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] raise self.value [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] updated_port = self._update_port( [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] _ensure_no_port_binding_failure(port) [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] raise exception.PortBindingFailed(port_id=port['id']) [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] nova.exception.PortBindingFailed: Binding failed for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e, please check neutron logs for more information. [ 742.071930] env[62183]: ERROR nova.compute.manager [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] [ 742.072469] env[62183]: DEBUG nova.compute.utils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Binding failed for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.072469] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.685s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.072469] env[62183]: DEBUG nova.objects.instance [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62183) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 742.075080] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Build of instance 3cb4be60-b626-4cc9-91eb-2bd9b778f153 was re-scheduled: Binding failed for port f3e8d80a-3ff2-4a54-9885-5a7339cdc21e, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 742.075583] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 742.075823] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquiring lock "refresh_cache-3cb4be60-b626-4cc9-91eb-2bd9b778f153" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.075929] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Acquired lock "refresh_cache-3cb4be60-b626-4cc9-91eb-2bd9b778f153" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.076147] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.365200] env[62183]: INFO nova.compute.manager [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] [instance: a1cabc70-e6c0-4c9b-8302-6ed8805f2182] Took 1.03 seconds to deallocate network for instance. [ 742.537780] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 742.598832] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.675945] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.056204] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.084524] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e145cd3f-89d1-408c-982d-015bc09c71fa tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.085502] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.965s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.179413] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Releasing lock "refresh_cache-3cb4be60-b626-4cc9-91eb-2bd9b778f153" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.179712] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 743.179921] env[62183]: DEBUG nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.180152] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.204996] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.400908] env[62183]: INFO nova.scheduler.client.report [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Deleted allocations for instance a1cabc70-e6c0-4c9b-8302-6ed8805f2182 [ 743.518935] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.518935] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.707573] env[62183]: DEBUG nova.network.neutron [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.812557] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8d065d-c267-4ac8-ab69-b8e38f7ad551 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.819664] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb03ec2-7bfe-49e1-a77e-cf17cf80892c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.851034] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03657044-3e47-4888-9aff-f58628f4462f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.856938] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907a5640-26ee-4dfb-a333-1f9ed22e029a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.869449] env[62183]: DEBUG nova.compute.provider_tree [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.908832] env[62183]: DEBUG oslo_concurrency.lockutils [None req-616659ff-d98e-48cb-81b2-b9b16bc790a6 tempest-ServersAdminNegativeTestJSON-2037602512 tempest-ServersAdminNegativeTestJSON-2037602512-project-member] Lock "a1cabc70-e6c0-4c9b-8302-6ed8805f2182" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.419s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.212959] env[62183]: INFO nova.compute.manager [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] [instance: 3cb4be60-b626-4cc9-91eb-2bd9b778f153] Took 1.03 seconds to deallocate network for instance. [ 744.372682] env[62183]: DEBUG nova.scheduler.client.report [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 744.410574] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.877552] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.792s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.878216] env[62183]: ERROR nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875, please check neutron logs for more information. [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] Traceback (most recent call last): [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self.driver.spawn(context, instance, image_meta, [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] vm_ref = self.build_virtual_machine(instance, [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.878216] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] for vif in network_info: [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] return self._sync_wrapper(fn, *args, **kwargs) [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self.wait() [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self[:] = self._gt.wait() [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] return self._exit_event.wait() [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] current.throw(*self._exc) [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.878598] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] result = function(*args, **kwargs) [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] return func(*args, **kwargs) [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] raise e [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] nwinfo = self.network_api.allocate_for_instance( [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] created_port_ids = self._update_ports_for_instance( [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] with excutils.save_and_reraise_exception(): [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] self.force_reraise() [ 744.879026] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.879540] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] raise self.value [ 744.879540] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.879540] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] updated_port = self._update_port( [ 744.879540] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.879540] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] _ensure_no_port_binding_failure(port) [ 744.879540] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.879540] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] raise exception.PortBindingFailed(port_id=port['id']) [ 744.879540] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] nova.exception.PortBindingFailed: Binding failed for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875, please check neutron logs for more information. [ 744.879540] env[62183]: ERROR nova.compute.manager [instance: f25ef564-b649-4557-9847-b5d994079d20] [ 744.879540] env[62183]: DEBUG nova.compute.utils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Binding failed for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 744.880324] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.308s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.881752] env[62183]: INFO nova.compute.claims [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.884440] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Build of instance f25ef564-b649-4557-9847-b5d994079d20 was re-scheduled: Binding failed for port 2d47ffe3-d54f-4d24-a23d-c50c46cf3875, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 744.884901] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 744.885659] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-f25ef564-b649-4557-9847-b5d994079d20" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.885659] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-f25ef564-b649-4557-9847-b5d994079d20" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.885659] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.931960] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.273940] env[62183]: INFO nova.scheduler.client.report [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Deleted allocations for instance 3cb4be60-b626-4cc9-91eb-2bd9b778f153 [ 745.411115] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.542725] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.783659] env[62183]: DEBUG oslo_concurrency.lockutils [None req-47da1bd0-e36b-49a2-a006-c576ce289168 tempest-SecurityGroupsTestJSON-601988892 tempest-SecurityGroupsTestJSON-601988892-project-member] Lock "3cb4be60-b626-4cc9-91eb-2bd9b778f153" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.952s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.045308] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-f25ef564-b649-4557-9847-b5d994079d20" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.045599] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.045715] env[62183]: DEBUG nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.045926] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.065500] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.157696] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4cc02a-24db-4d4d-b3b1-2821cf2baa8d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.166062] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc829bad-c4d9-4c0b-a550-5e521f608d17 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.198615] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d5b2b8-f31c-446c-9cb9-a7d9073883d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.206118] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c0beca-624d-4d45-98d0-fd05e0dd6182 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.219719] env[62183]: DEBUG nova.compute.provider_tree [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.286388] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.569489] env[62183]: DEBUG nova.network.neutron [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.723261] env[62183]: DEBUG nova.scheduler.client.report [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 746.812547] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.074729] env[62183]: INFO nova.compute.manager [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: f25ef564-b649-4557-9847-b5d994079d20] Took 1.03 seconds to deallocate network for instance. [ 747.232863] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.233638] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.269s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.234861] env[62183]: DEBUG nova.objects.instance [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lazy-loading 'resources' on Instance uuid c2182423-e7fe-4ae2-be94-e88a1e49bab9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 747.738108] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Acquiring lock "e7607725-3179-49a3-80bf-b86a1a2fa6f6" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.738399] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Lock "e7607725-3179-49a3-80bf-b86a1a2fa6f6" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.072469] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62854e0b-2112-41b9-a961-84bdedfe3e28 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.082846] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8e69b6-ba31-4605-9b87-8335fccdf573 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.122783] env[62183]: INFO nova.scheduler.client.report [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Deleted allocations for instance f25ef564-b649-4557-9847-b5d994079d20 [ 748.131215] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310bbf75-2e76-45e2-83c5-5bee1181d69f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.142887] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958b0ef2-a223-45e3-bc04-b7230339425b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.159379] env[62183]: DEBUG nova.compute.provider_tree [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.245433] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Lock "e7607725-3179-49a3-80bf-b86a1a2fa6f6" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.507s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.245972] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.631664] env[62183]: DEBUG oslo_concurrency.lockutils [None req-35024d80-f822-41f6-9c10-203018493461 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "f25ef564-b649-4557-9847-b5d994079d20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.271s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.663645] env[62183]: DEBUG nova.scheduler.client.report [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 748.751631] env[62183]: DEBUG nova.compute.utils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.753107] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.753293] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 748.801268] env[62183]: DEBUG nova.policy [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ba669768de1942379033e0de4692aa95', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a6716e6a9b8949ccbb3fed8dc39d2d8b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 749.111928] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "7465df88-c68c-49d3-9a91-6fff0d06957b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.112300] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7465df88-c68c-49d3-9a91-6fff0d06957b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.137539] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.169540] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.936s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.173460] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.731s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.189150] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Successfully created port: 28bd6733-3506-4a7f-9686-b80c6fa14c7a {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.212024] env[62183]: INFO nova.scheduler.client.report [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Deleted allocations for instance c2182423-e7fe-4ae2-be94-e88a1e49bab9 [ 749.256745] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.674159] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.723046] env[62183]: DEBUG oslo_concurrency.lockutils [None req-203c5503-b5c6-49a4-bd35-adcff49476cc tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "c2182423-e7fe-4ae2-be94-e88a1e49bab9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.465s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.000252] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5521c5a0-a020-40e4-b0cf-c6c906c6fb36 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.010614] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061ba92e-c546-43ea-ad61-ccb5d5cb2379 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.045784] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6250eda7-e68f-4f48-b68c-df75ef470f67 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.054483] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96fd411c-0205-45bc-9b88-b43f299f305d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.063112] env[62183]: DEBUG nova.compute.manager [req-663a9f49-7ba7-4b40-9806-0e0f26ddd238 req-f0bd826a-23e2-4e83-8cba-5f6152188071 service nova] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Received event network-changed-28bd6733-3506-4a7f-9686-b80c6fa14c7a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.063950] env[62183]: DEBUG nova.compute.manager [req-663a9f49-7ba7-4b40-9806-0e0f26ddd238 req-f0bd826a-23e2-4e83-8cba-5f6152188071 service nova] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Refreshing instance network info cache due to event network-changed-28bd6733-3506-4a7f-9686-b80c6fa14c7a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 750.063950] env[62183]: DEBUG oslo_concurrency.lockutils [req-663a9f49-7ba7-4b40-9806-0e0f26ddd238 req-f0bd826a-23e2-4e83-8cba-5f6152188071 service nova] Acquiring lock "refresh_cache-4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.063950] env[62183]: DEBUG oslo_concurrency.lockutils [req-663a9f49-7ba7-4b40-9806-0e0f26ddd238 req-f0bd826a-23e2-4e83-8cba-5f6152188071 service nova] Acquired lock "refresh_cache-4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.063950] env[62183]: DEBUG nova.network.neutron [req-663a9f49-7ba7-4b40-9806-0e0f26ddd238 req-f0bd826a-23e2-4e83-8cba-5f6152188071 service nova] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Refreshing network info cache for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 750.075711] env[62183]: DEBUG nova.compute.provider_tree [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.129919] env[62183]: DEBUG nova.network.neutron [req-663a9f49-7ba7-4b40-9806-0e0f26ddd238 req-f0bd826a-23e2-4e83-8cba-5f6152188071 service nova] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.138195] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "4af88a37-5aa2-47af-9dd9-8233b1bbf077" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.138434] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "4af88a37-5aa2-47af-9dd9-8233b1bbf077" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.138642] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "4af88a37-5aa2-47af-9dd9-8233b1bbf077-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.138822] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "4af88a37-5aa2-47af-9dd9-8233b1bbf077-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.138987] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "4af88a37-5aa2-47af-9dd9-8233b1bbf077-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.144022] env[62183]: INFO nova.compute.manager [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Terminating instance [ 750.144022] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "refresh_cache-4af88a37-5aa2-47af-9dd9-8233b1bbf077" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.144022] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquired lock "refresh_cache-4af88a37-5aa2-47af-9dd9-8233b1bbf077" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.144022] env[62183]: DEBUG nova.network.neutron [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.266063] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.270168] env[62183]: DEBUG nova.network.neutron [req-663a9f49-7ba7-4b40-9806-0e0f26ddd238 req-f0bd826a-23e2-4e83-8cba-5f6152188071 service nova] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.305800] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.306045] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.306230] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.306422] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.306562] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.306706] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.306928] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.307116] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.307351] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.307532] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.307803] env[62183]: DEBUG nova.virt.hardware [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.308849] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ab250d-0be9-4136-9cca-dc40a6e53827 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.317647] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d24525-8abe-41ed-bfe9-0966db04cbd3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.391277] env[62183]: ERROR nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a, please check neutron logs for more information. [ 750.391277] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.391277] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.391277] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.391277] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.391277] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.391277] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.391277] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.391277] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.391277] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 750.391277] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.391277] env[62183]: ERROR nova.compute.manager raise self.value [ 750.391277] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.391277] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.391277] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.391277] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.391829] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.391829] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.391829] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a, please check neutron logs for more information. [ 750.391829] env[62183]: ERROR nova.compute.manager [ 750.391829] env[62183]: Traceback (most recent call last): [ 750.391829] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.391829] env[62183]: listener.cb(fileno) [ 750.391829] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.391829] env[62183]: result = function(*args, **kwargs) [ 750.391829] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.391829] env[62183]: return func(*args, **kwargs) [ 750.391829] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.391829] env[62183]: raise e [ 750.391829] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.391829] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 750.391829] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.391829] env[62183]: created_port_ids = self._update_ports_for_instance( [ 750.391829] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.391829] env[62183]: with excutils.save_and_reraise_exception(): [ 750.391829] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.391829] env[62183]: self.force_reraise() [ 750.391829] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.391829] env[62183]: raise self.value [ 750.391829] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.391829] env[62183]: updated_port = self._update_port( [ 750.391829] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.391829] env[62183]: _ensure_no_port_binding_failure(port) [ 750.391829] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.391829] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.392669] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a, please check neutron logs for more information. [ 750.392669] env[62183]: Removing descriptor: 14 [ 750.392669] env[62183]: ERROR nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a, please check neutron logs for more information. [ 750.392669] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Traceback (most recent call last): [ 750.392669] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.392669] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] yield resources [ 750.392669] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.392669] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self.driver.spawn(context, instance, image_meta, [ 750.392669] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 750.392669] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.392669] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.392669] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] vm_ref = self.build_virtual_machine(instance, [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] for vif in network_info: [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] return self._sync_wrapper(fn, *args, **kwargs) [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self.wait() [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self[:] = self._gt.wait() [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] return self._exit_event.wait() [ 750.393031] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] result = hub.switch() [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] return self.greenlet.switch() [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] result = function(*args, **kwargs) [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] return func(*args, **kwargs) [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] raise e [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] nwinfo = self.network_api.allocate_for_instance( [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.393510] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] created_port_ids = self._update_ports_for_instance( [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] with excutils.save_and_reraise_exception(): [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self.force_reraise() [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] raise self.value [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] updated_port = self._update_port( [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] _ensure_no_port_binding_failure(port) [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.393914] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] raise exception.PortBindingFailed(port_id=port['id']) [ 750.394329] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] nova.exception.PortBindingFailed: Binding failed for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a, please check neutron logs for more information. [ 750.394329] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] [ 750.394329] env[62183]: INFO nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Terminating instance [ 750.395232] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Acquiring lock "refresh_cache-4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.584255] env[62183]: DEBUG nova.scheduler.client.report [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 750.668100] env[62183]: DEBUG nova.network.neutron [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.764966] env[62183]: DEBUG nova.network.neutron [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.773355] env[62183]: DEBUG oslo_concurrency.lockutils [req-663a9f49-7ba7-4b40-9806-0e0f26ddd238 req-f0bd826a-23e2-4e83-8cba-5f6152188071 service nova] Releasing lock "refresh_cache-4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.773355] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Acquired lock "refresh_cache-4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.773355] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.094376] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.921s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.094376] env[62183]: ERROR nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5cdeefba-3915-4199-9d8f-2518dc131d2e, please check neutron logs for more information. [ 751.094376] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Traceback (most recent call last): [ 751.094376] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.094376] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self.driver.spawn(context, instance, image_meta, [ 751.094376] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 751.094376] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.094376] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.094376] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] vm_ref = self.build_virtual_machine(instance, [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] for vif in network_info: [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] return self._sync_wrapper(fn, *args, **kwargs) [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self.wait() [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self[:] = self._gt.wait() [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] return self._exit_event.wait() [ 751.094780] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] current.throw(*self._exc) [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] result = function(*args, **kwargs) [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] return func(*args, **kwargs) [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] raise e [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] nwinfo = self.network_api.allocate_for_instance( [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] created_port_ids = self._update_ports_for_instance( [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.095194] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] with excutils.save_and_reraise_exception(): [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] self.force_reraise() [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] raise self.value [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] updated_port = self._update_port( [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] _ensure_no_port_binding_failure(port) [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] raise exception.PortBindingFailed(port_id=port['id']) [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] nova.exception.PortBindingFailed: Binding failed for port 5cdeefba-3915-4199-9d8f-2518dc131d2e, please check neutron logs for more information. [ 751.095575] env[62183]: ERROR nova.compute.manager [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] [ 751.096172] env[62183]: DEBUG nova.compute.utils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Binding failed for port 5cdeefba-3915-4199-9d8f-2518dc131d2e, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.097272] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.456s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.101036] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Build of instance b0761551-b2a3-4a38-a32d-f94a36c47d94 was re-scheduled: Binding failed for port 5cdeefba-3915-4199-9d8f-2518dc131d2e, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 751.102257] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 751.102546] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Acquiring lock "refresh_cache-b0761551-b2a3-4a38-a32d-f94a36c47d94" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.105221] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Acquired lock "refresh_cache-b0761551-b2a3-4a38-a32d-f94a36c47d94" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.105401] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.269733] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Releasing lock "refresh_cache-4af88a37-5aa2-47af-9dd9-8233b1bbf077" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.270225] env[62183]: DEBUG nova.compute.manager [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.270421] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.271350] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52eeba0-6eb2-4614-b3ce-bc7c4e93c87f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.281210] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 751.281452] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9f0b8ca-8078-40b8-93c0-d99a66e1c6d3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.287682] env[62183]: DEBUG oslo_vmware.api [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 751.287682] env[62183]: value = "task-1386956" [ 751.287682] env[62183]: _type = "Task" [ 751.287682] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.294893] env[62183]: DEBUG oslo_vmware.api [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386956, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.304419] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.433018] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.628554] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.648865] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.649375] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.773509] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.799141] env[62183]: DEBUG oslo_vmware.api [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386956, 'name': PowerOffVM_Task, 'duration_secs': 0.131182} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.801761] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 751.801966] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 751.802603] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4efd7d5-60f8-4c53-a77e-983d9fe63c28 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.827571] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 751.827780] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 751.827957] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Deleting the datastore file [datastore1] 4af88a37-5aa2-47af-9dd9-8233b1bbf077 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 751.828228] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b2f5cce-3f30-401b-9e8d-33d0276b1da0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.835595] env[62183]: DEBUG oslo_vmware.api [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for the task: (returnval){ [ 751.835595] env[62183]: value = "task-1386958" [ 751.835595] env[62183]: _type = "Task" [ 751.835595] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.845302] env[62183]: DEBUG oslo_vmware.api [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.885995] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc68f1bc-a10f-4267-9ee3-11b71aa0023c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.896021] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb936a4-c826-4ec2-9fbb-e153e60be5f2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.935742] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Releasing lock "refresh_cache-4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.936452] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.936747] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.937591] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f96eabe-8546-4c2d-8be0-675642d6c235 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.940247] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33e9dbfb-07a5-44f9-aeba-4e50f47f80be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.948698] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d50afe5-2d35-4919-b86d-9ee370a466e5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.954792] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef6d9d2-ad90-4e05-90bf-ced38b6cf237 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.974951] env[62183]: DEBUG nova.compute.provider_tree [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.980195] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f could not be found. [ 751.980453] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.980676] env[62183]: INFO nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 751.980945] env[62183]: DEBUG oslo.service.loopingcall [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.981386] env[62183]: DEBUG nova.compute.manager [-] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.981490] env[62183]: DEBUG nova.network.neutron [-] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.000820] env[62183]: DEBUG nova.network.neutron [-] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.094182] env[62183]: DEBUG nova.compute.manager [req-1f4a3891-03f1-4a04-a57b-99cff66c85f1 req-3ba599d9-5a64-4f3e-ae71-3356682b17c6 service nova] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Received event network-vif-deleted-28bd6733-3506-4a7f-9686-b80c6fa14c7a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.278151] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Releasing lock "refresh_cache-b0761551-b2a3-4a38-a32d-f94a36c47d94" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.278395] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 752.278543] env[62183]: DEBUG nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.278725] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.295725] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.345858] env[62183]: DEBUG oslo_vmware.api [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.482348] env[62183]: DEBUG nova.scheduler.client.report [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 752.504245] env[62183]: DEBUG nova.network.neutron [-] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.798016] env[62183]: DEBUG nova.network.neutron [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.847831] env[62183]: DEBUG oslo_vmware.api [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.989627] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.891s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.989627] env[62183]: ERROR nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8739dad5-b146-432a-b96b-6ef9a12276d6, please check neutron logs for more information. [ 752.989627] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Traceback (most recent call last): [ 752.989627] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.989627] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self.driver.spawn(context, instance, image_meta, [ 752.989627] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 752.989627] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.989627] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.989627] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] vm_ref = self.build_virtual_machine(instance, [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] for vif in network_info: [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] return self._sync_wrapper(fn, *args, **kwargs) [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self.wait() [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self[:] = self._gt.wait() [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] return self._exit_event.wait() [ 752.990049] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] result = hub.switch() [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] return self.greenlet.switch() [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] result = function(*args, **kwargs) [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] return func(*args, **kwargs) [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] raise e [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] nwinfo = self.network_api.allocate_for_instance( [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 752.990435] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] created_port_ids = self._update_ports_for_instance( [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] with excutils.save_and_reraise_exception(): [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] self.force_reraise() [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] raise self.value [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] updated_port = self._update_port( [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] _ensure_no_port_binding_failure(port) [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.990787] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] raise exception.PortBindingFailed(port_id=port['id']) [ 752.991130] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] nova.exception.PortBindingFailed: Binding failed for port 8739dad5-b146-432a-b96b-6ef9a12276d6, please check neutron logs for more information. [ 752.991130] env[62183]: ERROR nova.compute.manager [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] [ 752.991130] env[62183]: DEBUG nova.compute.utils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Binding failed for port 8739dad5-b146-432a-b96b-6ef9a12276d6, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.991130] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.456s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.992322] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Build of instance 430dcd52-17ce-4710-b9c8-39c28410e356 was re-scheduled: Binding failed for port 8739dad5-b146-432a-b96b-6ef9a12276d6, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 752.992892] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 752.992978] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Acquiring lock "refresh_cache-430dcd52-17ce-4710-b9c8-39c28410e356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.995337] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Acquired lock "refresh_cache-430dcd52-17ce-4710-b9c8-39c28410e356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.995337] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.005923] env[62183]: INFO nova.compute.manager [-] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Took 1.02 seconds to deallocate network for instance. [ 753.008178] env[62183]: DEBUG nova.compute.claims [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 753.008364] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.300702] env[62183]: INFO nova.compute.manager [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] [instance: b0761551-b2a3-4a38-a32d-f94a36c47d94] Took 1.02 seconds to deallocate network for instance. [ 753.348415] env[62183]: DEBUG oslo_vmware.api [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.513293] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.574224] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.733843] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquiring lock "68e791b4-61db-4b6b-a30a-ccf5d6657643" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.734089] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "68e791b4-61db-4b6b-a30a-ccf5d6657643" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.848248] env[62183]: DEBUG oslo_vmware.api [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Task: {'id': task-1386958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.690244} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.849119] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 753.849189] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 753.849363] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 753.849535] env[62183]: INFO nova.compute.manager [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Took 2.58 seconds to destroy the instance on the hypervisor. [ 753.849762] env[62183]: DEBUG oslo.service.loopingcall [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.849944] env[62183]: DEBUG nova.compute.manager [-] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.850061] env[62183]: DEBUG nova.network.neutron [-] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 753.864673] env[62183]: DEBUG nova.network.neutron [-] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.019587] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 4af88a37-5aa2-47af-9dd9-8233b1bbf077 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 754.079017] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Releasing lock "refresh_cache-430dcd52-17ce-4710-b9c8-39c28410e356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.079017] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 754.079017] env[62183]: DEBUG nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.079017] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.092502] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.335849] env[62183]: INFO nova.scheduler.client.report [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Deleted allocations for instance b0761551-b2a3-4a38-a32d-f94a36c47d94 [ 754.367561] env[62183]: DEBUG nova.network.neutron [-] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.522914] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance b0761551-b2a3-4a38-a32d-f94a36c47d94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.595715] env[62183]: DEBUG nova.network.neutron [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.846054] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dfefbd62-a6f7-4b40-a11d-85cbacc193f8 tempest-ServersV294TestFqdnHostnames-1990801715 tempest-ServersV294TestFqdnHostnames-1990801715-project-member] Lock "b0761551-b2a3-4a38-a32d-f94a36c47d94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.969s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.872443] env[62183]: INFO nova.compute.manager [-] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Took 1.02 seconds to deallocate network for instance. [ 755.026747] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 430dcd52-17ce-4710-b9c8-39c28410e356 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.026908] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 45adf688-561c-49c0-8395-a16bf4e43740 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.027054] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance f6e8b9ae-a3ea-4164-9a04-a7bb240303b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.027180] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 755.098093] env[62183]: INFO nova.compute.manager [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] [instance: 430dcd52-17ce-4710-b9c8-39c28410e356] Took 1.02 seconds to deallocate network for instance. [ 755.349144] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.380396] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.530148] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 9278c981-042e-41a2-a4f8-d42a8e6fac39 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.872469] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.033066] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 3b231f56-5127-4ff9-9031-06bfd59a57d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.129627] env[62183]: INFO nova.scheduler.client.report [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Deleted allocations for instance 430dcd52-17ce-4710-b9c8-39c28410e356 [ 756.536045] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 3fef2148-244d-4575-8e3a-93bf70c7dd7c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.641614] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95283205-57c7-4fd6-8d6f-b3c876b6783a tempest-ServerExternalEventsTest-2013548673 tempest-ServerExternalEventsTest-2013548673-project-member] Lock "430dcd52-17ce-4710-b9c8-39c28410e356" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.417s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.039887] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 48909506-5abd-4dfd-9aac-f3c1ef0f8250 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.144826] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.542657] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 90be0169-9414-431b-9aa3-82d89bea4bb2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.667294] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.045879] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance fe04dc9a-9fda-41e8-b62e-78dc4c026968 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.548667] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance bd5c642b-05e2-4c96-8e22-54b02fc93263 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.052014] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 27ba0765-152b-4985-927d-99670818a0b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.555015] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 5a559b13-b1e0-414a-896a-e70648ed93d9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.057592] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 07899178-b53c-4b0e-877d-11774286212f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.560972] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 0dd90b88-ed92-4331-98c5-c7481ac1ae15 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.064152] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.566899] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 7465df88-c68c-49d3-9a91-6fff0d06957b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.924663] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquiring lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.924897] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.070089] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance ac555ffc-ce4e-4650-97fd-c26a3246fe4b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.576254] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 68e791b4-61db-4b6b-a30a-ccf5d6657643 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.576254] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 762.576254] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 762.799965] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d95c0a-9b1e-4551-b844-80e0465d3136 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.807873] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7c8246-d46e-49cd-9e7d-251fb6284e74 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.838746] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d6a573-ff4a-4ba7-944a-4536069c3fa0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.845890] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0886003-85e2-4395-8669-72a411be169b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.858626] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.363041] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 763.868711] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 763.869064] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.878s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.869284] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.997s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.584878] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c55cf2-6e24-4540-bc4c-de5feb13314f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.592415] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d224d4cd-5515-467f-8e97-61d018de86aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.621669] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae25ce22-5b2d-4dc9-adb6-ff2146a44999 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.628248] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e9e082-b807-4942-b1d6-7982a5837a44 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.640797] env[62183]: DEBUG nova.compute.provider_tree [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.143781] env[62183]: DEBUG nova.scheduler.client.report [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 765.146895] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.147060] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Cleaning up deleted instances {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 765.650343] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.781s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.651029] env[62183]: ERROR nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e4527402-f953-4a68-bd4e-d5fff9c83737, please check neutron logs for more information. [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Traceback (most recent call last): [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self.driver.spawn(context, instance, image_meta, [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] vm_ref = self.build_virtual_machine(instance, [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.651029] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] for vif in network_info: [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] return self._sync_wrapper(fn, *args, **kwargs) [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self.wait() [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self[:] = self._gt.wait() [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] return self._exit_event.wait() [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] result = hub.switch() [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 765.651522] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] return self.greenlet.switch() [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] result = function(*args, **kwargs) [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] return func(*args, **kwargs) [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] raise e [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] nwinfo = self.network_api.allocate_for_instance( [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] created_port_ids = self._update_ports_for_instance( [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] with excutils.save_and_reraise_exception(): [ 765.651951] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] self.force_reraise() [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] raise self.value [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] updated_port = self._update_port( [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] _ensure_no_port_binding_failure(port) [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] raise exception.PortBindingFailed(port_id=port['id']) [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] nova.exception.PortBindingFailed: Binding failed for port e4527402-f953-4a68-bd4e-d5fff9c83737, please check neutron logs for more information. [ 765.652416] env[62183]: ERROR nova.compute.manager [instance: 45adf688-561c-49c0-8395-a16bf4e43740] [ 765.652782] env[62183]: DEBUG nova.compute.utils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Binding failed for port e4527402-f953-4a68-bd4e-d5fff9c83737, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 765.653241] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.800s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.654672] env[62183]: INFO nova.compute.claims [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.659034] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] There are 4 instances to clean {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 765.659034] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: c2182423-e7fe-4ae2-be94-e88a1e49bab9] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 765.659969] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Build of instance 45adf688-561c-49c0-8395-a16bf4e43740 was re-scheduled: Binding failed for port e4527402-f953-4a68-bd4e-d5fff9c83737, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 765.660428] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 765.660648] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Acquiring lock "refresh_cache-45adf688-561c-49c0-8395-a16bf4e43740" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.660793] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Acquired lock "refresh_cache-45adf688-561c-49c0-8395-a16bf4e43740" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.660951] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.164935] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 8397279a-7a43-4fb0-bc27-17e17e258527] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 766.205893] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.329368] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.672158] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a43e24b0-c0f1-4e02-894d-0d2c8ed884d1] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 766.833036] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Releasing lock "refresh_cache-45adf688-561c-49c0-8395-a16bf4e43740" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.833036] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 766.833036] env[62183]: DEBUG nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.833036] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.849094] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.867472] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594f7ae3-4e91-4cb3-afc6-3bb95575e7b1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.874792] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a7615b-dde7-4a37-95f3-5c66b651d66f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.905357] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c904926-8611-43e4-a9a1-32dc394adb6a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.912543] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded5f7bf-139e-4b30-9787-8514e8f4765e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.925503] env[62183]: DEBUG nova.compute.provider_tree [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.175832] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a11cfd43-7516-4f98-8484-9ba2a0c5276d] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 767.352557] env[62183]: DEBUG nova.network.neutron [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.428380] env[62183]: DEBUG nova.scheduler.client.report [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 767.682734] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.682910] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Cleaning up deleted instances with incomplete migration {{(pid=62183) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 767.855260] env[62183]: INFO nova.compute.manager [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] [instance: 45adf688-561c-49c0-8395-a16bf4e43740] Took 1.02 seconds to deallocate network for instance. [ 767.933136] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.933705] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 767.936187] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.532s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.185445] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.441025] env[62183]: DEBUG nova.compute.utils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.445479] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.445646] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 768.497726] env[62183]: DEBUG nova.policy [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe9a6a94d3d34140b5a8db5bee644dfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '56583cefe64e4af698ad1e608695aada', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 768.667374] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c13ca1-bc57-42c4-8b35-23c1b9adc3b0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.675854] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300deb40-f918-43e4-9695-caccbdf586e8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.707085] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ae4ed4-110e-40b9-8bd3-970bf671892f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.714385] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f6fe7e-ef89-43ec-bf57-0252ccd8391f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.727759] env[62183]: DEBUG nova.compute.provider_tree [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.784418] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Successfully created port: 9a85cc55-729b-45e3-8d44-7a322b6217cf {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.886784] env[62183]: INFO nova.scheduler.client.report [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Deleted allocations for instance 45adf688-561c-49c0-8395-a16bf4e43740 [ 768.948875] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.231563] env[62183]: DEBUG nova.scheduler.client.report [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 769.395119] env[62183]: DEBUG oslo_concurrency.lockutils [None req-67780ba9-ff03-4aa9-b754-a8032c629772 tempest-ServersTestJSON-2144542796 tempest-ServersTestJSON-2144542796-project-member] Lock "45adf688-561c-49c0-8395-a16bf4e43740" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.858s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.572342] env[62183]: DEBUG nova.compute.manager [req-1c4e680e-4b0e-4fe5-9780-63c13cfd5ff8 req-1a25343a-9915-4f85-a409-73125c4abe79 service nova] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Received event network-changed-9a85cc55-729b-45e3-8d44-7a322b6217cf {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.572663] env[62183]: DEBUG nova.compute.manager [req-1c4e680e-4b0e-4fe5-9780-63c13cfd5ff8 req-1a25343a-9915-4f85-a409-73125c4abe79 service nova] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Refreshing instance network info cache due to event network-changed-9a85cc55-729b-45e3-8d44-7a322b6217cf. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.572923] env[62183]: DEBUG oslo_concurrency.lockutils [req-1c4e680e-4b0e-4fe5-9780-63c13cfd5ff8 req-1a25343a-9915-4f85-a409-73125c4abe79 service nova] Acquiring lock "refresh_cache-9278c981-042e-41a2-a4f8-d42a8e6fac39" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.573121] env[62183]: DEBUG oslo_concurrency.lockutils [req-1c4e680e-4b0e-4fe5-9780-63c13cfd5ff8 req-1a25343a-9915-4f85-a409-73125c4abe79 service nova] Acquired lock "refresh_cache-9278c981-042e-41a2-a4f8-d42a8e6fac39" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.573327] env[62183]: DEBUG nova.network.neutron [req-1c4e680e-4b0e-4fe5-9780-63c13cfd5ff8 req-1a25343a-9915-4f85-a409-73125c4abe79 service nova] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Refreshing network info cache for port 9a85cc55-729b-45e3-8d44-7a322b6217cf {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 769.736266] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.800s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.736909] env[62183]: ERROR nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e18a9c5e-d140-412b-8d31-f0329a8b8375, please check neutron logs for more information. [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Traceback (most recent call last): [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self.driver.spawn(context, instance, image_meta, [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] vm_ref = self.build_virtual_machine(instance, [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.736909] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] for vif in network_info: [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] return self._sync_wrapper(fn, *args, **kwargs) [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self.wait() [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self[:] = self._gt.wait() [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] return self._exit_event.wait() [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] current.throw(*self._exc) [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.737301] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] result = function(*args, **kwargs) [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] return func(*args, **kwargs) [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] raise e [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] nwinfo = self.network_api.allocate_for_instance( [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] created_port_ids = self._update_ports_for_instance( [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] with excutils.save_and_reraise_exception(): [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] self.force_reraise() [ 769.737733] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.738168] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] raise self.value [ 769.738168] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.738168] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] updated_port = self._update_port( [ 769.738168] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.738168] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] _ensure_no_port_binding_failure(port) [ 769.738168] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.738168] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] raise exception.PortBindingFailed(port_id=port['id']) [ 769.738168] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] nova.exception.PortBindingFailed: Binding failed for port e18a9c5e-d140-412b-8d31-f0329a8b8375, please check neutron logs for more information. [ 769.738168] env[62183]: ERROR nova.compute.manager [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] [ 769.738168] env[62183]: DEBUG nova.compute.utils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Binding failed for port e18a9c5e-d140-412b-8d31-f0329a8b8375, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.740249] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.684s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.741479] env[62183]: INFO nova.compute.claims [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.747018] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Build of instance f6e8b9ae-a3ea-4164-9a04-a7bb240303b8 was re-scheduled: Binding failed for port e18a9c5e-d140-412b-8d31-f0329a8b8375, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 769.747018] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 769.747018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "refresh_cache-f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.747018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "refresh_cache-f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.747018] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.895278] env[62183]: ERROR nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9a85cc55-729b-45e3-8d44-7a322b6217cf, please check neutron logs for more information. [ 769.895278] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 769.895278] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.895278] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 769.895278] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.895278] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 769.895278] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.895278] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 769.895278] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.895278] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 769.895278] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.895278] env[62183]: ERROR nova.compute.manager raise self.value [ 769.895278] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.895278] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 769.895278] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.895278] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 769.895885] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.895885] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 769.895885] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9a85cc55-729b-45e3-8d44-7a322b6217cf, please check neutron logs for more information. [ 769.895885] env[62183]: ERROR nova.compute.manager [ 769.895885] env[62183]: Traceback (most recent call last): [ 769.895885] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 769.895885] env[62183]: listener.cb(fileno) [ 769.895885] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.895885] env[62183]: result = function(*args, **kwargs) [ 769.895885] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.895885] env[62183]: return func(*args, **kwargs) [ 769.895885] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.895885] env[62183]: raise e [ 769.895885] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.895885] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 769.895885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.895885] env[62183]: created_port_ids = self._update_ports_for_instance( [ 769.895885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.895885] env[62183]: with excutils.save_and_reraise_exception(): [ 769.895885] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.895885] env[62183]: self.force_reraise() [ 769.895885] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.895885] env[62183]: raise self.value [ 769.895885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.895885] env[62183]: updated_port = self._update_port( [ 769.895885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.895885] env[62183]: _ensure_no_port_binding_failure(port) [ 769.895885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.895885] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 769.896889] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 9a85cc55-729b-45e3-8d44-7a322b6217cf, please check neutron logs for more information. [ 769.896889] env[62183]: Removing descriptor: 14 [ 769.901049] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.958182] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.002349] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.002646] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.002815] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.002995] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.003162] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.003312] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.003518] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.003670] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.003830] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.003985] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.004269] env[62183]: DEBUG nova.virt.hardware [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.005508] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8696feb8-553f-4ead-91b8-187bb4b018fb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.013800] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9147b5-3cf5-4dff-8433-aaac3b4b9700 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.027149] env[62183]: ERROR nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9a85cc55-729b-45e3-8d44-7a322b6217cf, please check neutron logs for more information. [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Traceback (most recent call last): [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] yield resources [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self.driver.spawn(context, instance, image_meta, [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] vm_ref = self.build_virtual_machine(instance, [ 770.027149] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] for vif in network_info: [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] return self._sync_wrapper(fn, *args, **kwargs) [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self.wait() [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self[:] = self._gt.wait() [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] return self._exit_event.wait() [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 770.027594] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] current.throw(*self._exc) [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] result = function(*args, **kwargs) [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] return func(*args, **kwargs) [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] raise e [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] nwinfo = self.network_api.allocate_for_instance( [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] created_port_ids = self._update_ports_for_instance( [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] with excutils.save_and_reraise_exception(): [ 770.028040] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self.force_reraise() [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] raise self.value [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] updated_port = self._update_port( [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] _ensure_no_port_binding_failure(port) [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] raise exception.PortBindingFailed(port_id=port['id']) [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] nova.exception.PortBindingFailed: Binding failed for port 9a85cc55-729b-45e3-8d44-7a322b6217cf, please check neutron logs for more information. [ 770.028496] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] [ 770.028496] env[62183]: INFO nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Terminating instance [ 770.030063] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "refresh_cache-9278c981-042e-41a2-a4f8-d42a8e6fac39" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.095122] env[62183]: DEBUG nova.network.neutron [req-1c4e680e-4b0e-4fe5-9780-63c13cfd5ff8 req-1a25343a-9915-4f85-a409-73125c4abe79 service nova] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.238137] env[62183]: DEBUG nova.network.neutron [req-1c4e680e-4b0e-4fe5-9780-63c13cfd5ff8 req-1a25343a-9915-4f85-a409-73125c4abe79 service nova] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.286609] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.426520] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.432488] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.706029] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.706258] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.707191] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 770.707191] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Rebuilding the list of instances to heal {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 770.740390] env[62183]: DEBUG oslo_concurrency.lockutils [req-1c4e680e-4b0e-4fe5-9780-63c13cfd5ff8 req-1a25343a-9915-4f85-a409-73125c4abe79 service nova] Releasing lock "refresh_cache-9278c981-042e-41a2-a4f8-d42a8e6fac39" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.741502] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquired lock "refresh_cache-9278c981-042e-41a2-a4f8-d42a8e6fac39" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.741771] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.937264] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "refresh_cache-f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.937501] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 770.937677] env[62183]: DEBUG nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.937848] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.965886] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.994090] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f41770-da3f-48a4-8f87-4a4c30d2f56b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.001942] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d902b6-9f93-449a-abb2-8f6cdabf2b29 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.032873] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd981cbd-bf88-43b2-8791-daeedbcc542a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.040158] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad263049-5d10-4af2-843b-8624c557b5f7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.053769] env[62183]: DEBUG nova.compute.provider_tree [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.211638] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 771.212130] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 771.212130] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 771.230947] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "refresh_cache-4af88a37-5aa2-47af-9dd9-8233b1bbf077" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.231178] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquired lock "refresh_cache-4af88a37-5aa2-47af-9dd9-8233b1bbf077" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.231338] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Forcefully refreshing network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 771.231496] env[62183]: DEBUG nova.objects.instance [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lazy-loading 'info_cache' on Instance uuid 4af88a37-5aa2-47af-9dd9-8233b1bbf077 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 771.267880] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.407079] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.472912] env[62183]: DEBUG nova.network.neutron [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.556868] env[62183]: DEBUG nova.scheduler.client.report [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 771.605424] env[62183]: DEBUG nova.compute.manager [req-291d4f24-6696-488d-82e6-bbf3d17ab72f req-b15b76a1-f7d1-4542-8752-ea0571c43886 service nova] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Received event network-vif-deleted-9a85cc55-729b-45e3-8d44-7a322b6217cf {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.909653] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Releasing lock "refresh_cache-9278c981-042e-41a2-a4f8-d42a8e6fac39" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.910145] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.910356] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.911094] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21f28d72-c609-4b8d-94f6-f5bf599a801c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.920567] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce11de43-c2c7-4c3a-92f6-befdf99f4156 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.944034] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9278c981-042e-41a2-a4f8-d42a8e6fac39 could not be found. [ 771.944301] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.944517] env[62183]: INFO nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Took 0.03 seconds to destroy the instance on the hypervisor. [ 771.944782] env[62183]: DEBUG oslo.service.loopingcall [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.945021] env[62183]: DEBUG nova.compute.manager [-] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.945125] env[62183]: DEBUG nova.network.neutron [-] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.960946] env[62183]: DEBUG nova.network.neutron [-] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.975111] env[62183]: INFO nova.compute.manager [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: f6e8b9ae-a3ea-4164-9a04-a7bb240303b8] Took 1.04 seconds to deallocate network for instance. [ 772.061415] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.061949] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 772.064515] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.133s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.065900] env[62183]: INFO nova.compute.claims [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.266584] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.464106] env[62183]: DEBUG nova.network.neutron [-] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.570305] env[62183]: DEBUG nova.compute.utils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.574021] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 772.574191] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 772.626938] env[62183]: DEBUG nova.policy [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe9a6a94d3d34140b5a8db5bee644dfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '56583cefe64e4af698ad1e608695aada', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.850995] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.973801] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Successfully created port: f5db769a-3499-47e8-9d21-7ada9e155260 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.973801] env[62183]: INFO nova.compute.manager [-] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Took 1.02 seconds to deallocate network for instance. [ 772.974112] env[62183]: DEBUG nova.compute.claims [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 772.974282] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.020106] env[62183]: INFO nova.scheduler.client.report [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleted allocations for instance f6e8b9ae-a3ea-4164-9a04-a7bb240303b8 [ 773.075338] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 773.345199] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-206ce290-dfde-41c7-a92d-00131bdac3be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.353733] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b10526b-911f-421c-a573-5468e682ef97 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.358326] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Releasing lock "refresh_cache-4af88a37-5aa2-47af-9dd9-8233b1bbf077" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.358535] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Updated the network info_cache for instance {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 773.358734] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.397367] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.398220] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.399086] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95899ad-9ae5-4ef6-b6b2-453cc36419ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.402124] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.402310] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.403073] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.403247] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 773.403488] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.408348] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7bb028-7c20-4244-af51-fc4ee8d72c84 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.424324] env[62183]: DEBUG nova.compute.provider_tree [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.534356] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5d6ade86-d0a3-47e6-ab5c-81721396afdf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "f6e8b9ae-a3ea-4164-9a04-a7bb240303b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.229s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.807125] env[62183]: DEBUG nova.compute.manager [req-e82d3380-6dcd-42cb-9ba6-97923597fc86 req-fdd022ba-6b5d-4ef7-b179-6fbfcba86168 service nova] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Received event network-changed-f5db769a-3499-47e8-9d21-7ada9e155260 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.807125] env[62183]: DEBUG nova.compute.manager [req-e82d3380-6dcd-42cb-9ba6-97923597fc86 req-fdd022ba-6b5d-4ef7-b179-6fbfcba86168 service nova] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Refreshing instance network info cache due to event network-changed-f5db769a-3499-47e8-9d21-7ada9e155260. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.807125] env[62183]: DEBUG oslo_concurrency.lockutils [req-e82d3380-6dcd-42cb-9ba6-97923597fc86 req-fdd022ba-6b5d-4ef7-b179-6fbfcba86168 service nova] Acquiring lock "refresh_cache-3b231f56-5127-4ff9-9031-06bfd59a57d0" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.807125] env[62183]: DEBUG oslo_concurrency.lockutils [req-e82d3380-6dcd-42cb-9ba6-97923597fc86 req-fdd022ba-6b5d-4ef7-b179-6fbfcba86168 service nova] Acquired lock "refresh_cache-3b231f56-5127-4ff9-9031-06bfd59a57d0" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.807675] env[62183]: DEBUG nova.network.neutron [req-e82d3380-6dcd-42cb-9ba6-97923597fc86 req-fdd022ba-6b5d-4ef7-b179-6fbfcba86168 service nova] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Refreshing network info cache for port f5db769a-3499-47e8-9d21-7ada9e155260 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 773.906482] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.933191] env[62183]: DEBUG nova.scheduler.client.report [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 773.966271] env[62183]: ERROR nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f5db769a-3499-47e8-9d21-7ada9e155260, please check neutron logs for more information. [ 773.966271] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 773.966271] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.966271] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 773.966271] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.966271] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 773.966271] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.966271] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 773.966271] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.966271] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 773.966271] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.966271] env[62183]: ERROR nova.compute.manager raise self.value [ 773.966271] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.966271] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 773.966271] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.966271] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 773.966983] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.966983] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 773.966983] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f5db769a-3499-47e8-9d21-7ada9e155260, please check neutron logs for more information. [ 773.966983] env[62183]: ERROR nova.compute.manager [ 773.966983] env[62183]: Traceback (most recent call last): [ 773.966983] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 773.966983] env[62183]: listener.cb(fileno) [ 773.966983] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.966983] env[62183]: result = function(*args, **kwargs) [ 773.966983] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.966983] env[62183]: return func(*args, **kwargs) [ 773.966983] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.966983] env[62183]: raise e [ 773.966983] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.966983] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 773.966983] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.966983] env[62183]: created_port_ids = self._update_ports_for_instance( [ 773.966983] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.966983] env[62183]: with excutils.save_and_reraise_exception(): [ 773.966983] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.966983] env[62183]: self.force_reraise() [ 773.966983] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.966983] env[62183]: raise self.value [ 773.966983] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.966983] env[62183]: updated_port = self._update_port( [ 773.966983] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.966983] env[62183]: _ensure_no_port_binding_failure(port) [ 773.966983] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.966983] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 773.967626] env[62183]: nova.exception.PortBindingFailed: Binding failed for port f5db769a-3499-47e8-9d21-7ada9e155260, please check neutron logs for more information. [ 773.967626] env[62183]: Removing descriptor: 14 [ 774.037425] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 774.092019] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 774.123023] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.123023] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.123023] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.123637] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.123912] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.124199] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.127025] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.127025] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.127025] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.127025] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.127025] env[62183]: DEBUG nova.virt.hardware [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.127218] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3e5df7-4fbb-4ba3-8e9a-ebdafc9ffd35 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.137335] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd5c731-831e-4687-bf87-997639af5c20 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.153603] env[62183]: ERROR nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f5db769a-3499-47e8-9d21-7ada9e155260, please check neutron logs for more information. [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Traceback (most recent call last): [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] yield resources [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self.driver.spawn(context, instance, image_meta, [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] vm_ref = self.build_virtual_machine(instance, [ 774.153603] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] for vif in network_info: [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] return self._sync_wrapper(fn, *args, **kwargs) [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self.wait() [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self[:] = self._gt.wait() [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] return self._exit_event.wait() [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 774.154063] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] current.throw(*self._exc) [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] result = function(*args, **kwargs) [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] return func(*args, **kwargs) [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] raise e [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] nwinfo = self.network_api.allocate_for_instance( [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] created_port_ids = self._update_ports_for_instance( [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] with excutils.save_and_reraise_exception(): [ 774.155212] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self.force_reraise() [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] raise self.value [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] updated_port = self._update_port( [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] _ensure_no_port_binding_failure(port) [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] raise exception.PortBindingFailed(port_id=port['id']) [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] nova.exception.PortBindingFailed: Binding failed for port f5db769a-3499-47e8-9d21-7ada9e155260, please check neutron logs for more information. [ 774.155582] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] [ 774.155582] env[62183]: INFO nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Terminating instance [ 774.157173] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "refresh_cache-3b231f56-5127-4ff9-9031-06bfd59a57d0" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.339715] env[62183]: DEBUG nova.network.neutron [req-e82d3380-6dcd-42cb-9ba6-97923597fc86 req-fdd022ba-6b5d-4ef7-b179-6fbfcba86168 service nova] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.435747] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.436320] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.438855] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.626s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.440259] env[62183]: INFO nova.compute.claims [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.473356] env[62183]: DEBUG nova.network.neutron [req-e82d3380-6dcd-42cb-9ba6-97923597fc86 req-fdd022ba-6b5d-4ef7-b179-6fbfcba86168 service nova] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.565306] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.946043] env[62183]: DEBUG nova.compute.utils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.948580] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.949176] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 775.755710] env[62183]: DEBUG oslo_concurrency.lockutils [req-e82d3380-6dcd-42cb-9ba6-97923597fc86 req-fdd022ba-6b5d-4ef7-b179-6fbfcba86168 service nova] Releasing lock "refresh_cache-3b231f56-5127-4ff9-9031-06bfd59a57d0" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.756493] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.763344] env[62183]: DEBUG nova.policy [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '01f26bd7fa58489cb86c396785dd1062', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42ef7c4301bc4c4cb964bb2df1c19045', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 775.764673] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquired lock "refresh_cache-3b231f56-5127-4ff9-9031-06bfd59a57d0" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.764889] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.769380] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.769589] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.924972] env[62183]: DEBUG nova.compute.manager [req-c291f8a7-9613-48f8-8667-161035a0cf0b req-258a453f-8a95-493f-bff3-a908fdddb472 service nova] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Received event network-vif-deleted-f5db769a-3499-47e8-9d21-7ada9e155260 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.094503] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Successfully created port: c7364122-69c2-4c96-8086-736513d9265e {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.125271] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb85c20-56c9-4080-ad00-d6dff260ee6a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.133231] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6906af0f-cc85-41b0-9702-184aa19c6d32 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.162335] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee64e28d-6e08-4ec1-9f18-9a723aba0720 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.169942] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67eeca04-a63b-4d07-92fd-0f0eba395510 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.186982] env[62183]: DEBUG nova.compute.provider_tree [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.336924] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.462605] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.695365] env[62183]: DEBUG nova.scheduler.client.report [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 776.778420] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.797022] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.797022] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.808517] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.808758] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.808913] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.809107] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.809254] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.809400] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.809634] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.809815] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.809980] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.810192] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.810371] env[62183]: DEBUG nova.virt.hardware [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.811979] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f2c9e7-ec5a-41ce-882f-5a72c9b9c8de {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.820547] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbbc2a0-43b8-4ef1-860b-fae8ec5669d7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.965544] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Releasing lock "refresh_cache-3b231f56-5127-4ff9-9031-06bfd59a57d0" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.966078] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 776.966292] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 776.966683] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb56876d-b903-4f2d-9527-cf4c4e0f6421 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.975891] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25925e2c-7909-4300-a044-0b0d8ca61537 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.998249] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3b231f56-5127-4ff9-9031-06bfd59a57d0 could not be found. [ 776.998468] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.998653] env[62183]: INFO nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 776.999411] env[62183]: DEBUG oslo.service.loopingcall [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.999411] env[62183]: DEBUG nova.compute.manager [-] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.999411] env[62183]: DEBUG nova.network.neutron [-] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 777.016503] env[62183]: DEBUG nova.network.neutron [-] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.205030] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.763s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.205030] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.205727] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.532s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.207986] env[62183]: INFO nova.compute.claims [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.353106] env[62183]: ERROR nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c7364122-69c2-4c96-8086-736513d9265e, please check neutron logs for more information. [ 777.353106] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 777.353106] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.353106] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 777.353106] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.353106] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 777.353106] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.353106] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 777.353106] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.353106] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 777.353106] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.353106] env[62183]: ERROR nova.compute.manager raise self.value [ 777.353106] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.353106] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 777.353106] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.353106] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 777.353542] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.353542] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 777.353542] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c7364122-69c2-4c96-8086-736513d9265e, please check neutron logs for more information. [ 777.353542] env[62183]: ERROR nova.compute.manager [ 777.353542] env[62183]: Traceback (most recent call last): [ 777.353542] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 777.353542] env[62183]: listener.cb(fileno) [ 777.353542] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.353542] env[62183]: result = function(*args, **kwargs) [ 777.353542] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 777.353542] env[62183]: return func(*args, **kwargs) [ 777.353542] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.353542] env[62183]: raise e [ 777.353542] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.353542] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 777.353542] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.353542] env[62183]: created_port_ids = self._update_ports_for_instance( [ 777.353542] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.353542] env[62183]: with excutils.save_and_reraise_exception(): [ 777.353542] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.353542] env[62183]: self.force_reraise() [ 777.353542] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.353542] env[62183]: raise self.value [ 777.353542] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.353542] env[62183]: updated_port = self._update_port( [ 777.353542] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.353542] env[62183]: _ensure_no_port_binding_failure(port) [ 777.353542] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.353542] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 777.354247] env[62183]: nova.exception.PortBindingFailed: Binding failed for port c7364122-69c2-4c96-8086-736513d9265e, please check neutron logs for more information. [ 777.354247] env[62183]: Removing descriptor: 14 [ 777.354247] env[62183]: ERROR nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c7364122-69c2-4c96-8086-736513d9265e, please check neutron logs for more information. [ 777.354247] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Traceback (most recent call last): [ 777.354247] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 777.354247] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] yield resources [ 777.354247] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.354247] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self.driver.spawn(context, instance, image_meta, [ 777.354247] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 777.354247] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.354247] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.354247] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] vm_ref = self.build_virtual_machine(instance, [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] for vif in network_info: [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] return self._sync_wrapper(fn, *args, **kwargs) [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self.wait() [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self[:] = self._gt.wait() [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] return self._exit_event.wait() [ 777.354560] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] result = hub.switch() [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] return self.greenlet.switch() [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] result = function(*args, **kwargs) [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] return func(*args, **kwargs) [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] raise e [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] nwinfo = self.network_api.allocate_for_instance( [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.354881] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] created_port_ids = self._update_ports_for_instance( [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] with excutils.save_and_reraise_exception(): [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self.force_reraise() [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] raise self.value [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] updated_port = self._update_port( [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] _ensure_no_port_binding_failure(port) [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.355206] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] raise exception.PortBindingFailed(port_id=port['id']) [ 777.355501] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] nova.exception.PortBindingFailed: Binding failed for port c7364122-69c2-4c96-8086-736513d9265e, please check neutron logs for more information. [ 777.355501] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] [ 777.355501] env[62183]: INFO nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Terminating instance [ 777.358767] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Acquiring lock "refresh_cache-3fef2148-244d-4575-8e3a-93bf70c7dd7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.358767] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Acquired lock "refresh_cache-3fef2148-244d-4575-8e3a-93bf70c7dd7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.358767] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.518779] env[62183]: DEBUG nova.network.neutron [-] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.715039] env[62183]: DEBUG nova.compute.utils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 777.718242] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 777.718610] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 777.782715] env[62183]: DEBUG nova.policy [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f80808d61aa42f283d2d47631137403', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee3e2b1fecd74debb8362622a78039a2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 777.887211] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.963731] env[62183]: DEBUG nova.compute.manager [req-2bdf9bc2-a894-4965-ab98-86ca6712489d req-67288103-978f-40d4-8e52-7d159c7704aa service nova] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Received event network-changed-c7364122-69c2-4c96-8086-736513d9265e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.963927] env[62183]: DEBUG nova.compute.manager [req-2bdf9bc2-a894-4965-ab98-86ca6712489d req-67288103-978f-40d4-8e52-7d159c7704aa service nova] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Refreshing instance network info cache due to event network-changed-c7364122-69c2-4c96-8086-736513d9265e. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 777.964115] env[62183]: DEBUG oslo_concurrency.lockutils [req-2bdf9bc2-a894-4965-ab98-86ca6712489d req-67288103-978f-40d4-8e52-7d159c7704aa service nova] Acquiring lock "refresh_cache-3fef2148-244d-4575-8e3a-93bf70c7dd7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.021287] env[62183]: INFO nova.compute.manager [-] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Took 1.02 seconds to deallocate network for instance. [ 778.023664] env[62183]: DEBUG nova.compute.claims [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 778.023859] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.038215] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.119874] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Successfully created port: 54fe49a1-3468-438a-86c6-129f68209fa1 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.217603] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 778.493385] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36506d57-c428-4d06-a82a-23a9aef9d152 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.501559] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2767d058-2892-40ec-af58-eb0956ad8c83 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.531486] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b7bdf2-003d-41b7-91eb-4a5cee75682a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.538778] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae0a1b4-a3f8-455a-a969-049c147c1f1c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.544552] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Releasing lock "refresh_cache-3fef2148-244d-4575-8e3a-93bf70c7dd7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.545027] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 778.545275] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 778.545611] env[62183]: DEBUG oslo_concurrency.lockutils [req-2bdf9bc2-a894-4965-ab98-86ca6712489d req-67288103-978f-40d4-8e52-7d159c7704aa service nova] Acquired lock "refresh_cache-3fef2148-244d-4575-8e3a-93bf70c7dd7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.545804] env[62183]: DEBUG nova.network.neutron [req-2bdf9bc2-a894-4965-ab98-86ca6712489d req-67288103-978f-40d4-8e52-7d159c7704aa service nova] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Refreshing network info cache for port c7364122-69c2-4c96-8086-736513d9265e {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 778.546832] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4ad8e3e-412f-4122-a914-c50d18b94a46 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.557848] env[62183]: DEBUG nova.compute.provider_tree [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.563986] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2667c558-68f2-4f07-b990-60d8ffb09d3c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.587540] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3fef2148-244d-4575-8e3a-93bf70c7dd7c could not be found. [ 778.587767] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 778.587944] env[62183]: INFO nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 778.588201] env[62183]: DEBUG oslo.service.loopingcall [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.588442] env[62183]: DEBUG nova.compute.manager [-] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.588513] env[62183]: DEBUG nova.network.neutron [-] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.604784] env[62183]: DEBUG nova.network.neutron [-] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.049620] env[62183]: ERROR nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 54fe49a1-3468-438a-86c6-129f68209fa1, please check neutron logs for more information. [ 779.049620] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 779.049620] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.049620] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 779.049620] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.049620] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 779.049620] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.049620] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 779.049620] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.049620] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 779.049620] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.049620] env[62183]: ERROR nova.compute.manager raise self.value [ 779.049620] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.049620] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 779.049620] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.049620] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 779.050523] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.050523] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 779.050523] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 54fe49a1-3468-438a-86c6-129f68209fa1, please check neutron logs for more information. [ 779.050523] env[62183]: ERROR nova.compute.manager [ 779.050523] env[62183]: Traceback (most recent call last): [ 779.050523] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 779.050523] env[62183]: listener.cb(fileno) [ 779.050523] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.050523] env[62183]: result = function(*args, **kwargs) [ 779.050523] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.050523] env[62183]: return func(*args, **kwargs) [ 779.050523] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.050523] env[62183]: raise e [ 779.050523] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.050523] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 779.050523] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.050523] env[62183]: created_port_ids = self._update_ports_for_instance( [ 779.050523] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.050523] env[62183]: with excutils.save_and_reraise_exception(): [ 779.050523] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.050523] env[62183]: self.force_reraise() [ 779.050523] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.050523] env[62183]: raise self.value [ 779.050523] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.050523] env[62183]: updated_port = self._update_port( [ 779.050523] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.050523] env[62183]: _ensure_no_port_binding_failure(port) [ 779.050523] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.050523] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 779.051843] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 54fe49a1-3468-438a-86c6-129f68209fa1, please check neutron logs for more information. [ 779.051843] env[62183]: Removing descriptor: 14 [ 779.063167] env[62183]: DEBUG nova.scheduler.client.report [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 779.066839] env[62183]: DEBUG nova.network.neutron [req-2bdf9bc2-a894-4965-ab98-86ca6712489d req-67288103-978f-40d4-8e52-7d159c7704aa service nova] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.106730] env[62183]: DEBUG nova.network.neutron [-] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.153740] env[62183]: DEBUG nova.network.neutron [req-2bdf9bc2-a894-4965-ab98-86ca6712489d req-67288103-978f-40d4-8e52-7d159c7704aa service nova] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.231044] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.267384] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.267663] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.267892] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.268114] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.268267] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.268412] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.268621] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.268774] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.268935] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.269111] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.269282] env[62183]: DEBUG nova.virt.hardware [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.270134] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ec6c53-d40f-4bf9-8754-55900fb02904 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.277793] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9e40d5-face-4816-9e3e-a5403894a18e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.291379] env[62183]: ERROR nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 54fe49a1-3468-438a-86c6-129f68209fa1, please check neutron logs for more information. [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Traceback (most recent call last): [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] yield resources [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self.driver.spawn(context, instance, image_meta, [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] vm_ref = self.build_virtual_machine(instance, [ 779.291379] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] for vif in network_info: [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] return self._sync_wrapper(fn, *args, **kwargs) [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self.wait() [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self[:] = self._gt.wait() [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] return self._exit_event.wait() [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 779.291742] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] current.throw(*self._exc) [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] result = function(*args, **kwargs) [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] return func(*args, **kwargs) [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] raise e [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] nwinfo = self.network_api.allocate_for_instance( [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] created_port_ids = self._update_ports_for_instance( [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] with excutils.save_and_reraise_exception(): [ 779.292086] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self.force_reraise() [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] raise self.value [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] updated_port = self._update_port( [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] _ensure_no_port_binding_failure(port) [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] raise exception.PortBindingFailed(port_id=port['id']) [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] nova.exception.PortBindingFailed: Binding failed for port 54fe49a1-3468-438a-86c6-129f68209fa1, please check neutron logs for more information. [ 779.292438] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] [ 779.292438] env[62183]: INFO nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Terminating instance [ 779.293717] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "refresh_cache-48909506-5abd-4dfd-9aac-f3c1ef0f8250" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.293874] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquired lock "refresh_cache-48909506-5abd-4dfd-9aac-f3c1ef0f8250" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.294055] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.569783] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.570354] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.573387] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.565s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.609331] env[62183]: INFO nova.compute.manager [-] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Took 1.02 seconds to deallocate network for instance. [ 779.611639] env[62183]: DEBUG nova.compute.claims [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 779.611866] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.655953] env[62183]: DEBUG oslo_concurrency.lockutils [req-2bdf9bc2-a894-4965-ab98-86ca6712489d req-67288103-978f-40d4-8e52-7d159c7704aa service nova] Releasing lock "refresh_cache-3fef2148-244d-4575-8e3a-93bf70c7dd7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.656229] env[62183]: DEBUG nova.compute.manager [req-2bdf9bc2-a894-4965-ab98-86ca6712489d req-67288103-978f-40d4-8e52-7d159c7704aa service nova] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Received event network-vif-deleted-c7364122-69c2-4c96-8086-736513d9265e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.810742] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.877539] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.015194] env[62183]: DEBUG nova.compute.manager [req-b7e499df-ad64-446e-b0ca-e36245f624c3 req-b1004f4d-6769-455c-812f-399928ed21d8 service nova] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Received event network-changed-54fe49a1-3468-438a-86c6-129f68209fa1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.015350] env[62183]: DEBUG nova.compute.manager [req-b7e499df-ad64-446e-b0ca-e36245f624c3 req-b1004f4d-6769-455c-812f-399928ed21d8 service nova] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Refreshing instance network info cache due to event network-changed-54fe49a1-3468-438a-86c6-129f68209fa1. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 780.015738] env[62183]: DEBUG oslo_concurrency.lockutils [req-b7e499df-ad64-446e-b0ca-e36245f624c3 req-b1004f4d-6769-455c-812f-399928ed21d8 service nova] Acquiring lock "refresh_cache-48909506-5abd-4dfd-9aac-f3c1ef0f8250" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.077864] env[62183]: DEBUG nova.compute.utils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.081930] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.082128] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 780.134929] env[62183]: DEBUG nova.policy [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de43bd5c7cb64826a090cdf0626e31e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'baa89de457ab46d49cea507aec5819bd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.308664] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285a6d4b-c537-4e32-8026-42c0570be119 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.316256] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f2b66f-2dcd-4d44-99b1-8015eb3cc601 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.346817] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2507da65-3357-4075-8ce8-0c0f35eed575 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.354440] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57810d45-9161-4d53-8c3c-58e1fa91d6e5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.367957] env[62183]: DEBUG nova.compute.provider_tree [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.380216] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Releasing lock "refresh_cache-48909506-5abd-4dfd-9aac-f3c1ef0f8250" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.380589] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 780.380779] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.381089] env[62183]: DEBUG oslo_concurrency.lockutils [req-b7e499df-ad64-446e-b0ca-e36245f624c3 req-b1004f4d-6769-455c-812f-399928ed21d8 service nova] Acquired lock "refresh_cache-48909506-5abd-4dfd-9aac-f3c1ef0f8250" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.381248] env[62183]: DEBUG nova.network.neutron [req-b7e499df-ad64-446e-b0ca-e36245f624c3 req-b1004f4d-6769-455c-812f-399928ed21d8 service nova] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Refreshing network info cache for port 54fe49a1-3468-438a-86c6-129f68209fa1 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 780.382385] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-991cb7c1-6edf-4d89-905e-a65d4afb4f1a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.392117] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e06f8a-ae38-48ca-bc1d-ee557d3282b9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.417491] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 48909506-5abd-4dfd-9aac-f3c1ef0f8250 could not be found. [ 780.417706] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 780.417923] env[62183]: INFO nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Took 0.04 seconds to destroy the instance on the hypervisor. [ 780.418149] env[62183]: DEBUG oslo.service.loopingcall [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.418351] env[62183]: DEBUG nova.compute.manager [-] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.418445] env[62183]: DEBUG nova.network.neutron [-] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.443840] env[62183]: DEBUG nova.network.neutron [-] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.445178] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Successfully created port: 2358598c-6738-475e-b6bf-5be07a6070a6 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.585348] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.871113] env[62183]: DEBUG nova.scheduler.client.report [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 780.904741] env[62183]: DEBUG nova.network.neutron [req-b7e499df-ad64-446e-b0ca-e36245f624c3 req-b1004f4d-6769-455c-812f-399928ed21d8 service nova] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.948116] env[62183]: DEBUG nova.network.neutron [-] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.982616] env[62183]: DEBUG nova.network.neutron [req-b7e499df-ad64-446e-b0ca-e36245f624c3 req-b1004f4d-6769-455c-812f-399928ed21d8 service nova] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.359204] env[62183]: ERROR nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2358598c-6738-475e-b6bf-5be07a6070a6, please check neutron logs for more information. [ 781.359204] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 781.359204] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.359204] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 781.359204] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.359204] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 781.359204] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.359204] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 781.359204] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.359204] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 781.359204] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.359204] env[62183]: ERROR nova.compute.manager raise self.value [ 781.359204] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.359204] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 781.359204] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.359204] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 781.359885] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.359885] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 781.359885] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2358598c-6738-475e-b6bf-5be07a6070a6, please check neutron logs for more information. [ 781.359885] env[62183]: ERROR nova.compute.manager [ 781.359885] env[62183]: Traceback (most recent call last): [ 781.359885] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 781.359885] env[62183]: listener.cb(fileno) [ 781.359885] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.359885] env[62183]: result = function(*args, **kwargs) [ 781.359885] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.359885] env[62183]: return func(*args, **kwargs) [ 781.359885] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.359885] env[62183]: raise e [ 781.359885] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.359885] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 781.359885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.359885] env[62183]: created_port_ids = self._update_ports_for_instance( [ 781.359885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.359885] env[62183]: with excutils.save_and_reraise_exception(): [ 781.359885] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.359885] env[62183]: self.force_reraise() [ 781.359885] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.359885] env[62183]: raise self.value [ 781.359885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.359885] env[62183]: updated_port = self._update_port( [ 781.359885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.359885] env[62183]: _ensure_no_port_binding_failure(port) [ 781.359885] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.359885] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 781.360706] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 2358598c-6738-475e-b6bf-5be07a6070a6, please check neutron logs for more information. [ 781.360706] env[62183]: Removing descriptor: 14 [ 781.376389] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.803s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.377066] env[62183]: ERROR nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a, please check neutron logs for more information. [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Traceback (most recent call last): [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self.driver.spawn(context, instance, image_meta, [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] vm_ref = self.build_virtual_machine(instance, [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.377066] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] for vif in network_info: [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] return self._sync_wrapper(fn, *args, **kwargs) [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self.wait() [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self[:] = self._gt.wait() [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] return self._exit_event.wait() [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] result = hub.switch() [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 781.377323] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] return self.greenlet.switch() [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] result = function(*args, **kwargs) [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] return func(*args, **kwargs) [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] raise e [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] nwinfo = self.network_api.allocate_for_instance( [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] created_port_ids = self._update_ports_for_instance( [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] with excutils.save_and_reraise_exception(): [ 781.377639] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] self.force_reraise() [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] raise self.value [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] updated_port = self._update_port( [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] _ensure_no_port_binding_failure(port) [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] raise exception.PortBindingFailed(port_id=port['id']) [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] nova.exception.PortBindingFailed: Binding failed for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a, please check neutron logs for more information. [ 781.377917] env[62183]: ERROR nova.compute.manager [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] [ 781.378194] env[62183]: DEBUG nova.compute.utils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Binding failed for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.378966] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.999s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.379267] env[62183]: DEBUG nova.objects.instance [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lazy-loading 'resources' on Instance uuid 4af88a37-5aa2-47af-9dd9-8233b1bbf077 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 781.380984] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Build of instance 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f was re-scheduled: Binding failed for port 28bd6733-3506-4a7f-9686-b80c6fa14c7a, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 781.381410] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 781.381637] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Acquiring lock "refresh_cache-4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.381786] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Acquired lock "refresh_cache-4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.381945] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.451258] env[62183]: INFO nova.compute.manager [-] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Took 1.03 seconds to deallocate network for instance. [ 781.453537] env[62183]: DEBUG nova.compute.claims [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 781.453743] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.484710] env[62183]: DEBUG oslo_concurrency.lockutils [req-b7e499df-ad64-446e-b0ca-e36245f624c3 req-b1004f4d-6769-455c-812f-399928ed21d8 service nova] Releasing lock "refresh_cache-48909506-5abd-4dfd-9aac-f3c1ef0f8250" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.484963] env[62183]: DEBUG nova.compute.manager [req-b7e499df-ad64-446e-b0ca-e36245f624c3 req-b1004f4d-6769-455c-812f-399928ed21d8 service nova] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Received event network-vif-deleted-54fe49a1-3468-438a-86c6-129f68209fa1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.599019] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.625356] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.625617] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.625772] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.625948] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.626106] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.626255] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.626457] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.626613] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.626781] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.626937] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.627116] env[62183]: DEBUG nova.virt.hardware [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.627969] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8fb6ce2-801c-482c-8b24-d306676706ae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.635850] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a325c9d7-7a80-4c2e-a883-4d48467ef6d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.649639] env[62183]: ERROR nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2358598c-6738-475e-b6bf-5be07a6070a6, please check neutron logs for more information. [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Traceback (most recent call last): [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] yield resources [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self.driver.spawn(context, instance, image_meta, [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] vm_ref = self.build_virtual_machine(instance, [ 781.649639] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] for vif in network_info: [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] return self._sync_wrapper(fn, *args, **kwargs) [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self.wait() [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self[:] = self._gt.wait() [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] return self._exit_event.wait() [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.649987] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] current.throw(*self._exc) [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] result = function(*args, **kwargs) [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] return func(*args, **kwargs) [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] raise e [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] nwinfo = self.network_api.allocate_for_instance( [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] created_port_ids = self._update_ports_for_instance( [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] with excutils.save_and_reraise_exception(): [ 781.650379] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self.force_reraise() [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] raise self.value [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] updated_port = self._update_port( [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] _ensure_no_port_binding_failure(port) [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] raise exception.PortBindingFailed(port_id=port['id']) [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] nova.exception.PortBindingFailed: Binding failed for port 2358598c-6738-475e-b6bf-5be07a6070a6, please check neutron logs for more information. [ 781.650763] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] [ 781.650763] env[62183]: INFO nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Terminating instance [ 781.651877] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Acquiring lock "refresh_cache-90be0169-9414-431b-9aa3-82d89bea4bb2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.652057] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Acquired lock "refresh_cache-90be0169-9414-431b-9aa3-82d89bea4bb2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.652228] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.902397] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.976773] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.062391] env[62183]: DEBUG nova.compute.manager [req-5a0c80a6-31ce-42a7-8cac-4bd65e5ebcec req-d9c6ba2c-13e0-478c-afd1-7e65c82a4088 service nova] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Received event network-changed-2358598c-6738-475e-b6bf-5be07a6070a6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.062623] env[62183]: DEBUG nova.compute.manager [req-5a0c80a6-31ce-42a7-8cac-4bd65e5ebcec req-d9c6ba2c-13e0-478c-afd1-7e65c82a4088 service nova] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Refreshing instance network info cache due to event network-changed-2358598c-6738-475e-b6bf-5be07a6070a6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.062795] env[62183]: DEBUG oslo_concurrency.lockutils [req-5a0c80a6-31ce-42a7-8cac-4bd65e5ebcec req-d9c6ba2c-13e0-478c-afd1-7e65c82a4088 service nova] Acquiring lock "refresh_cache-90be0169-9414-431b-9aa3-82d89bea4bb2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.103659] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e2dfca-4620-4f1c-8e81-d37d964bd629 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.110907] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c11dbd-2310-4c30-8f5e-b082fac05450 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.139979] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e43929-e0c1-49a7-b365-ce38f039bc9a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.147139] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc5a255-62fc-485a-ae67-ac7ee98f7e7a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.162059] env[62183]: DEBUG nova.compute.provider_tree [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.174844] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.265978] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.481463] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Releasing lock "refresh_cache-4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.481736] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 782.481736] env[62183]: DEBUG nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.484032] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.497035] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.664887] env[62183]: DEBUG nova.scheduler.client.report [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 782.768746] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Releasing lock "refresh_cache-90be0169-9414-431b-9aa3-82d89bea4bb2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.769233] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 782.769428] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 782.769760] env[62183]: DEBUG oslo_concurrency.lockutils [req-5a0c80a6-31ce-42a7-8cac-4bd65e5ebcec req-d9c6ba2c-13e0-478c-afd1-7e65c82a4088 service nova] Acquired lock "refresh_cache-90be0169-9414-431b-9aa3-82d89bea4bb2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.769934] env[62183]: DEBUG nova.network.neutron [req-5a0c80a6-31ce-42a7-8cac-4bd65e5ebcec req-d9c6ba2c-13e0-478c-afd1-7e65c82a4088 service nova] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Refreshing network info cache for port 2358598c-6738-475e-b6bf-5be07a6070a6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 782.771301] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e47ccf9d-cc65-4037-836b-ebb4c1369d3e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.780734] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10d8100-eb1f-4e03-b5e5-4f5f8624d647 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.803439] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 90be0169-9414-431b-9aa3-82d89bea4bb2 could not be found. [ 782.803671] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.803849] env[62183]: INFO nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 782.804097] env[62183]: DEBUG oslo.service.loopingcall [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.804304] env[62183]: DEBUG nova.compute.manager [-] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.804397] env[62183]: DEBUG nova.network.neutron [-] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.818559] env[62183]: DEBUG nova.network.neutron [-] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.000305] env[62183]: DEBUG nova.network.neutron [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.169653] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.172326] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.300s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.173818] env[62183]: INFO nova.compute.claims [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.189952] env[62183]: INFO nova.scheduler.client.report [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Deleted allocations for instance 4af88a37-5aa2-47af-9dd9-8233b1bbf077 [ 783.289111] env[62183]: DEBUG nova.network.neutron [req-5a0c80a6-31ce-42a7-8cac-4bd65e5ebcec req-d9c6ba2c-13e0-478c-afd1-7e65c82a4088 service nova] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.320757] env[62183]: DEBUG nova.network.neutron [-] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.367165] env[62183]: DEBUG nova.network.neutron [req-5a0c80a6-31ce-42a7-8cac-4bd65e5ebcec req-d9c6ba2c-13e0-478c-afd1-7e65c82a4088 service nova] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.503617] env[62183]: INFO nova.compute.manager [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] [instance: 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f] Took 1.02 seconds to deallocate network for instance. [ 783.697509] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1561e294-2d6a-4a57-98a0-35bcc4619eb5 tempest-ServerShowV247Test-411793433 tempest-ServerShowV247Test-411793433-project-member] Lock "4af88a37-5aa2-47af-9dd9-8233b1bbf077" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.559s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.823286] env[62183]: INFO nova.compute.manager [-] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Took 1.02 seconds to deallocate network for instance. [ 783.825478] env[62183]: DEBUG nova.compute.claims [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 783.825648] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.869724] env[62183]: DEBUG oslo_concurrency.lockutils [req-5a0c80a6-31ce-42a7-8cac-4bd65e5ebcec req-d9c6ba2c-13e0-478c-afd1-7e65c82a4088 service nova] Releasing lock "refresh_cache-90be0169-9414-431b-9aa3-82d89bea4bb2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.869961] env[62183]: DEBUG nova.compute.manager [req-5a0c80a6-31ce-42a7-8cac-4bd65e5ebcec req-d9c6ba2c-13e0-478c-afd1-7e65c82a4088 service nova] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Received event network-vif-deleted-2358598c-6738-475e-b6bf-5be07a6070a6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.421284] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1c125e-76af-4a47-bbab-9667fbccb876 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.429066] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eef7e4a-cfaf-452b-99c1-7abede30386d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.459263] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784b6e8e-352a-4b0b-8c1d-a32e83d84c4f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.466447] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e3ed05-7926-4943-a0bd-407a74cc77c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.479462] env[62183]: DEBUG nova.compute.provider_tree [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.533602] env[62183]: INFO nova.scheduler.client.report [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Deleted allocations for instance 4287ae36-4c16-4f9d-802a-51ef0dc1ad8f [ 784.985596] env[62183]: DEBUG nova.scheduler.client.report [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 785.041807] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2ce0dcbc-9c4b-4237-8b83-7f5991b85d19 tempest-ServerGroupTestJSON-235580742 tempest-ServerGroupTestJSON-235580742-project-member] Lock "4287ae36-4c16-4f9d-802a-51ef0dc1ad8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.287s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.491534] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.492097] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.495419] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.828s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.496791] env[62183]: INFO nova.compute.claims [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.544162] env[62183]: DEBUG nova.compute.manager [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 786.002828] env[62183]: DEBUG nova.compute.utils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.004230] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.004396] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 786.055298] env[62183]: DEBUG nova.policy [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1aeb0a71b54e444080dc88ed57a3c6e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1bdee19c4b9c41299a1591edc98f0f4b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.063383] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.306774] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Successfully created port: 39ddd09f-b5f7-419f-8415-69db4af4d169 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.507949] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.752926] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bf3959-cfcb-44f2-b098-01450b7240ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.761143] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd06632-c3ae-4411-b489-91ec70ccf2ea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.797835] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ce772b-f8a3-48ee-afb8-a5f12cb35e74 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.806077] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7c634c-f59d-468f-9064-e95117096452 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.819334] env[62183]: DEBUG nova.compute.provider_tree [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.325128] env[62183]: DEBUG nova.scheduler.client.report [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 787.334293] env[62183]: DEBUG nova.compute.manager [req-c302bde3-d47c-4863-98a4-123b02d522f2 req-ba13f2fc-fada-4224-87d6-8b64c8b48f63 service nova] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Received event network-changed-39ddd09f-b5f7-419f-8415-69db4af4d169 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.335082] env[62183]: DEBUG nova.compute.manager [req-c302bde3-d47c-4863-98a4-123b02d522f2 req-ba13f2fc-fada-4224-87d6-8b64c8b48f63 service nova] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Refreshing instance network info cache due to event network-changed-39ddd09f-b5f7-419f-8415-69db4af4d169. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 787.335249] env[62183]: DEBUG oslo_concurrency.lockutils [req-c302bde3-d47c-4863-98a4-123b02d522f2 req-ba13f2fc-fada-4224-87d6-8b64c8b48f63 service nova] Acquiring lock "refresh_cache-fe04dc9a-9fda-41e8-b62e-78dc4c026968" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.335283] env[62183]: DEBUG oslo_concurrency.lockutils [req-c302bde3-d47c-4863-98a4-123b02d522f2 req-ba13f2fc-fada-4224-87d6-8b64c8b48f63 service nova] Acquired lock "refresh_cache-fe04dc9a-9fda-41e8-b62e-78dc4c026968" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.336278] env[62183]: DEBUG nova.network.neutron [req-c302bde3-d47c-4863-98a4-123b02d522f2 req-ba13f2fc-fada-4224-87d6-8b64c8b48f63 service nova] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Refreshing network info cache for port 39ddd09f-b5f7-419f-8415-69db4af4d169 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 787.521354] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.524672] env[62183]: ERROR nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 39ddd09f-b5f7-419f-8415-69db4af4d169, please check neutron logs for more information. [ 787.524672] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 787.524672] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.524672] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 787.524672] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.524672] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 787.524672] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.524672] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 787.524672] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.524672] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 787.524672] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.524672] env[62183]: ERROR nova.compute.manager raise self.value [ 787.524672] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.524672] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 787.524672] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.524672] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 787.525076] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.525076] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 787.525076] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 39ddd09f-b5f7-419f-8415-69db4af4d169, please check neutron logs for more information. [ 787.525076] env[62183]: ERROR nova.compute.manager [ 787.525076] env[62183]: Traceback (most recent call last): [ 787.525076] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 787.525076] env[62183]: listener.cb(fileno) [ 787.525076] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.525076] env[62183]: result = function(*args, **kwargs) [ 787.525076] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 787.525076] env[62183]: return func(*args, **kwargs) [ 787.525076] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.525076] env[62183]: raise e [ 787.525076] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.525076] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 787.525076] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.525076] env[62183]: created_port_ids = self._update_ports_for_instance( [ 787.525076] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.525076] env[62183]: with excutils.save_and_reraise_exception(): [ 787.525076] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.525076] env[62183]: self.force_reraise() [ 787.525076] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.525076] env[62183]: raise self.value [ 787.525076] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.525076] env[62183]: updated_port = self._update_port( [ 787.525076] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.525076] env[62183]: _ensure_no_port_binding_failure(port) [ 787.525076] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.525076] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 787.525758] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 39ddd09f-b5f7-419f-8415-69db4af4d169, please check neutron logs for more information. [ 787.525758] env[62183]: Removing descriptor: 14 [ 787.551607] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.551859] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.552023] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.552210] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.552355] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.552498] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.552726] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.553635] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.553635] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.553635] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.553635] env[62183]: DEBUG nova.virt.hardware [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.555849] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ace8270-d0e4-4c4c-8e32-cb58f539f85a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.562573] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36962c3a-7997-40f2-8685-2e5f62602517 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.578629] env[62183]: ERROR nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 39ddd09f-b5f7-419f-8415-69db4af4d169, please check neutron logs for more information. [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Traceback (most recent call last): [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] yield resources [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self.driver.spawn(context, instance, image_meta, [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] vm_ref = self.build_virtual_machine(instance, [ 787.578629] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] for vif in network_info: [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] return self._sync_wrapper(fn, *args, **kwargs) [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self.wait() [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self[:] = self._gt.wait() [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] return self._exit_event.wait() [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 787.578953] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] current.throw(*self._exc) [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] result = function(*args, **kwargs) [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] return func(*args, **kwargs) [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] raise e [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] nwinfo = self.network_api.allocate_for_instance( [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] created_port_ids = self._update_ports_for_instance( [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] with excutils.save_and_reraise_exception(): [ 787.579257] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self.force_reraise() [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] raise self.value [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] updated_port = self._update_port( [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] _ensure_no_port_binding_failure(port) [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] raise exception.PortBindingFailed(port_id=port['id']) [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] nova.exception.PortBindingFailed: Binding failed for port 39ddd09f-b5f7-419f-8415-69db4af4d169, please check neutron logs for more information. [ 787.579532] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] [ 787.579532] env[62183]: INFO nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Terminating instance [ 787.581103] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Acquiring lock "refresh_cache-fe04dc9a-9fda-41e8-b62e-78dc4c026968" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.831566] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.336s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.832123] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.834734] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.408s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.836225] env[62183]: INFO nova.compute.claims [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.857027] env[62183]: DEBUG nova.network.neutron [req-c302bde3-d47c-4863-98a4-123b02d522f2 req-ba13f2fc-fada-4224-87d6-8b64c8b48f63 service nova] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.957522] env[62183]: DEBUG nova.network.neutron [req-c302bde3-d47c-4863-98a4-123b02d522f2 req-ba13f2fc-fada-4224-87d6-8b64c8b48f63 service nova] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.340839] env[62183]: DEBUG nova.compute.utils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.345127] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 788.345420] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 788.387415] env[62183]: DEBUG nova.policy [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89ad04efd23b40e6a3eea5b3e6fd021e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b61c6a99cbda435481bb72f20929f03f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.459710] env[62183]: DEBUG oslo_concurrency.lockutils [req-c302bde3-d47c-4863-98a4-123b02d522f2 req-ba13f2fc-fada-4224-87d6-8b64c8b48f63 service nova] Releasing lock "refresh_cache-fe04dc9a-9fda-41e8-b62e-78dc4c026968" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.460166] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Acquired lock "refresh_cache-fe04dc9a-9fda-41e8-b62e-78dc4c026968" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.460355] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.753126] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Successfully created port: 45b75c14-62b4-4d5e-be95-8e02039570c1 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.849195] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.984404] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.125793] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.128756] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa5e78c-c6ea-44eb-837e-3cdcb3c837bc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.138561] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e704848-da9e-4d0b-b348-bf9b6855926f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.171860] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1c5af4-d2a4-4c13-8990-77c2350dd207 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.180552] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6663656d-14de-4fa0-a745-3c468bd58e45 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.196688] env[62183]: DEBUG nova.compute.provider_tree [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.373978] env[62183]: DEBUG nova.compute.manager [req-12ca7afb-ac08-4c8a-8bf3-4df1b205903b req-a37dd80f-8aa3-4509-bec5-1abb0ab0a91f service nova] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Received event network-vif-deleted-39ddd09f-b5f7-419f-8415-69db4af4d169 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.632574] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Releasing lock "refresh_cache-fe04dc9a-9fda-41e8-b62e-78dc4c026968" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.633105] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 789.633339] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.634075] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-167feb88-6f0c-4efc-ad0b-534d414883fa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.643396] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63c4c33-2258-4ecd-9fd3-2d1d42afd71f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.666092] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fe04dc9a-9fda-41e8-b62e-78dc4c026968 could not be found. [ 789.666335] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.666514] env[62183]: INFO nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Took 0.03 seconds to destroy the instance on the hypervisor. [ 789.666754] env[62183]: DEBUG oslo.service.loopingcall [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.666999] env[62183]: DEBUG nova.compute.manager [-] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.667104] env[62183]: DEBUG nova.network.neutron [-] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 789.690044] env[62183]: DEBUG nova.network.neutron [-] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.699837] env[62183]: DEBUG nova.scheduler.client.report [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 789.788686] env[62183]: ERROR nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 45b75c14-62b4-4d5e-be95-8e02039570c1, please check neutron logs for more information. [ 789.788686] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 789.788686] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.788686] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 789.788686] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.788686] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 789.788686] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.788686] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 789.788686] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.788686] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 789.788686] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.788686] env[62183]: ERROR nova.compute.manager raise self.value [ 789.788686] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.788686] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 789.788686] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.788686] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 789.789121] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.789121] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 789.789121] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 45b75c14-62b4-4d5e-be95-8e02039570c1, please check neutron logs for more information. [ 789.789121] env[62183]: ERROR nova.compute.manager [ 789.789121] env[62183]: Traceback (most recent call last): [ 789.789121] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 789.789121] env[62183]: listener.cb(fileno) [ 789.789121] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.789121] env[62183]: result = function(*args, **kwargs) [ 789.789121] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.789121] env[62183]: return func(*args, **kwargs) [ 789.789121] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.789121] env[62183]: raise e [ 789.789121] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.789121] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 789.789121] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.789121] env[62183]: created_port_ids = self._update_ports_for_instance( [ 789.789121] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.789121] env[62183]: with excutils.save_and_reraise_exception(): [ 789.789121] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.789121] env[62183]: self.force_reraise() [ 789.789121] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.789121] env[62183]: raise self.value [ 789.789121] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.789121] env[62183]: updated_port = self._update_port( [ 789.789121] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.789121] env[62183]: _ensure_no_port_binding_failure(port) [ 789.789121] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.789121] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 789.790599] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 45b75c14-62b4-4d5e-be95-8e02039570c1, please check neutron logs for more information. [ 789.790599] env[62183]: Removing descriptor: 14 [ 789.862040] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.886380] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.886630] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.886784] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.886961] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.887119] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.887263] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.887465] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.887621] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.887784] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.887944] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.888131] env[62183]: DEBUG nova.virt.hardware [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.888959] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1aeec31-107d-48fd-ba4e-659aefa00a92 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.897362] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca2aba6-eea1-428c-9569-dd8fb98126b8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.910875] env[62183]: ERROR nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 45b75c14-62b4-4d5e-be95-8e02039570c1, please check neutron logs for more information. [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Traceback (most recent call last): [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] yield resources [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self.driver.spawn(context, instance, image_meta, [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self._vmops.spawn(context, instance, image_meta, injected_files, [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] vm_ref = self.build_virtual_machine(instance, [ 789.910875] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] vif_infos = vmwarevif.get_vif_info(self._session, [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] for vif in network_info: [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] return self._sync_wrapper(fn, *args, **kwargs) [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self.wait() [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self[:] = self._gt.wait() [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] return self._exit_event.wait() [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 789.911215] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] current.throw(*self._exc) [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] result = function(*args, **kwargs) [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] return func(*args, **kwargs) [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] raise e [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] nwinfo = self.network_api.allocate_for_instance( [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] created_port_ids = self._update_ports_for_instance( [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] with excutils.save_and_reraise_exception(): [ 789.911523] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self.force_reraise() [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] raise self.value [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] updated_port = self._update_port( [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] _ensure_no_port_binding_failure(port) [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] raise exception.PortBindingFailed(port_id=port['id']) [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] nova.exception.PortBindingFailed: Binding failed for port 45b75c14-62b4-4d5e-be95-8e02039570c1, please check neutron logs for more information. [ 789.911796] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] [ 789.911796] env[62183]: INFO nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Terminating instance [ 789.913473] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-bd5c642b-05e2-4c96-8e22-54b02fc93263" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.913632] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-bd5c642b-05e2-4c96-8e22-54b02fc93263" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.913803] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.192092] env[62183]: DEBUG nova.network.neutron [-] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.209205] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.209789] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 790.213009] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.239s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.431753] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.513718] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.695582] env[62183]: INFO nova.compute.manager [-] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Took 1.03 seconds to deallocate network for instance. [ 790.697873] env[62183]: DEBUG nova.compute.claims [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 790.698070] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.721298] env[62183]: DEBUG nova.compute.utils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.722783] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.722951] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 790.761929] env[62183]: DEBUG nova.policy [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e53beb806094cd98c694aaffdad3b99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e74d26cb8f5a47a48dc025b8058c5768', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.931996] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c13b69-eaa6-438f-b1df-83303a452ed9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.939081] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648c72ac-cfa5-4b0a-aae0-a065d57cb1db {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.968397] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a447de2-8f70-4778-8f00-cc962c89b815 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.975373] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd12f08-2fcf-4b43-a8e6-09675bc43f0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.988029] env[62183]: DEBUG nova.compute.provider_tree [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.016079] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-bd5c642b-05e2-4c96-8e22-54b02fc93263" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.016506] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 791.016696] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 791.016956] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f343c45-6dfc-4fd9-84e1-52108d0deb12 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.025113] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbfd9437-fb49-42d9-8826-454ca005e0f4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.047699] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bd5c642b-05e2-4c96-8e22-54b02fc93263 could not be found. [ 791.047891] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.048084] env[62183]: INFO nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Took 0.03 seconds to destroy the instance on the hypervisor. [ 791.048324] env[62183]: DEBUG oslo.service.loopingcall [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.048533] env[62183]: DEBUG nova.compute.manager [-] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.048627] env[62183]: DEBUG nova.network.neutron [-] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.069166] env[62183]: DEBUG nova.network.neutron [-] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.087690] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Successfully created port: ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 791.225723] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.402186] env[62183]: DEBUG nova.compute.manager [req-9c4d1a91-21c8-4c40-8023-5ab257591ead req-7b6d07e1-d3a4-4e73-ad8c-9c8bae4ed763 service nova] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Received event network-changed-45b75c14-62b4-4d5e-be95-8e02039570c1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.404435] env[62183]: DEBUG nova.compute.manager [req-9c4d1a91-21c8-4c40-8023-5ab257591ead req-7b6d07e1-d3a4-4e73-ad8c-9c8bae4ed763 service nova] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Refreshing instance network info cache due to event network-changed-45b75c14-62b4-4d5e-be95-8e02039570c1. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 791.404435] env[62183]: DEBUG oslo_concurrency.lockutils [req-9c4d1a91-21c8-4c40-8023-5ab257591ead req-7b6d07e1-d3a4-4e73-ad8c-9c8bae4ed763 service nova] Acquiring lock "refresh_cache-bd5c642b-05e2-4c96-8e22-54b02fc93263" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.404435] env[62183]: DEBUG oslo_concurrency.lockutils [req-9c4d1a91-21c8-4c40-8023-5ab257591ead req-7b6d07e1-d3a4-4e73-ad8c-9c8bae4ed763 service nova] Acquired lock "refresh_cache-bd5c642b-05e2-4c96-8e22-54b02fc93263" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.404435] env[62183]: DEBUG nova.network.neutron [req-9c4d1a91-21c8-4c40-8023-5ab257591ead req-7b6d07e1-d3a4-4e73-ad8c-9c8bae4ed763 service nova] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Refreshing network info cache for port 45b75c14-62b4-4d5e-be95-8e02039570c1 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 791.493715] env[62183]: DEBUG nova.scheduler.client.report [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 791.574334] env[62183]: DEBUG nova.network.neutron [-] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.928581] env[62183]: DEBUG nova.network.neutron [req-9c4d1a91-21c8-4c40-8023-5ab257591ead req-7b6d07e1-d3a4-4e73-ad8c-9c8bae4ed763 service nova] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.996323] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.783s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.997103] env[62183]: ERROR nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9a85cc55-729b-45e3-8d44-7a322b6217cf, please check neutron logs for more information. [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Traceback (most recent call last): [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self.driver.spawn(context, instance, image_meta, [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] vm_ref = self.build_virtual_machine(instance, [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.997103] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] for vif in network_info: [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] return self._sync_wrapper(fn, *args, **kwargs) [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self.wait() [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self[:] = self._gt.wait() [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] return self._exit_event.wait() [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] current.throw(*self._exc) [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.997847] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] result = function(*args, **kwargs) [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] return func(*args, **kwargs) [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] raise e [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] nwinfo = self.network_api.allocate_for_instance( [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] created_port_ids = self._update_ports_for_instance( [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] with excutils.save_and_reraise_exception(): [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] self.force_reraise() [ 791.998571] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.999192] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] raise self.value [ 791.999192] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.999192] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] updated_port = self._update_port( [ 791.999192] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.999192] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] _ensure_no_port_binding_failure(port) [ 791.999192] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.999192] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] raise exception.PortBindingFailed(port_id=port['id']) [ 791.999192] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] nova.exception.PortBindingFailed: Binding failed for port 9a85cc55-729b-45e3-8d44-7a322b6217cf, please check neutron logs for more information. [ 791.999192] env[62183]: ERROR nova.compute.manager [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] [ 791.999192] env[62183]: DEBUG nova.compute.utils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Binding failed for port 9a85cc55-729b-45e3-8d44-7a322b6217cf, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 791.999667] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.093s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.999667] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.999667] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 791.999667] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.434s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.001145] env[62183]: INFO nova.compute.claims [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.008796] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Build of instance 9278c981-042e-41a2-a4f8-d42a8e6fac39 was re-scheduled: Binding failed for port 9a85cc55-729b-45e3-8d44-7a322b6217cf, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 792.009272] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 792.009603] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "refresh_cache-9278c981-042e-41a2-a4f8-d42a8e6fac39" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.009699] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquired lock "refresh_cache-9278c981-042e-41a2-a4f8-d42a8e6fac39" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.009901] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.011877] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3a4d87-56f8-4f6c-9236-270aa5963f0f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.022961] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa8d2c6-5740-4a9d-a1fd-c230ed6969a8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.041812] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bcfb5b-0522-4ebc-a0be-7b192b371413 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.050111] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c368bb3a-968d-4345-b12c-6277ab0e4f7f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.054044] env[62183]: DEBUG nova.network.neutron [req-9c4d1a91-21c8-4c40-8023-5ab257591ead req-7b6d07e1-d3a4-4e73-ad8c-9c8bae4ed763 service nova] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.056633] env[62183]: ERROR nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6, please check neutron logs for more information. [ 792.056633] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 792.056633] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.056633] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 792.056633] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.056633] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 792.056633] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.056633] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 792.056633] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.056633] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 792.056633] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.056633] env[62183]: ERROR nova.compute.manager raise self.value [ 792.056633] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.056633] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 792.056633] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.056633] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 792.057276] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.057276] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 792.057276] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6, please check neutron logs for more information. [ 792.057276] env[62183]: ERROR nova.compute.manager [ 792.057276] env[62183]: Traceback (most recent call last): [ 792.057276] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 792.057276] env[62183]: listener.cb(fileno) [ 792.057276] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.057276] env[62183]: result = function(*args, **kwargs) [ 792.057276] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 792.057276] env[62183]: return func(*args, **kwargs) [ 792.057276] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.057276] env[62183]: raise e [ 792.057276] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.057276] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 792.057276] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.057276] env[62183]: created_port_ids = self._update_ports_for_instance( [ 792.057276] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.057276] env[62183]: with excutils.save_and_reraise_exception(): [ 792.057276] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.057276] env[62183]: self.force_reraise() [ 792.057276] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.057276] env[62183]: raise self.value [ 792.057276] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.057276] env[62183]: updated_port = self._update_port( [ 792.057276] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.057276] env[62183]: _ensure_no_port_binding_failure(port) [ 792.057276] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.057276] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 792.058371] env[62183]: nova.exception.PortBindingFailed: Binding failed for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6, please check neutron logs for more information. [ 792.058371] env[62183]: Removing descriptor: 14 [ 792.084816] env[62183]: INFO nova.compute.manager [-] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Took 1.04 seconds to deallocate network for instance. [ 792.085282] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181132MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 792.085500] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.087739] env[62183]: DEBUG nova.compute.claims [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 792.087911] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.239837] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.276020] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.276020] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.276020] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.276237] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.276237] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.276410] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.276774] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.277076] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.277365] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.277645] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.277925] env[62183]: DEBUG nova.virt.hardware [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.278870] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b20bc2-a705-4887-8b5e-b0d59d1554a1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.287095] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f35fb0-9128-4d3b-b20c-b3069bcb780e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.301669] env[62183]: ERROR nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6, please check neutron logs for more information. [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] Traceback (most recent call last): [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] yield resources [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self.driver.spawn(context, instance, image_meta, [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] vm_ref = self.build_virtual_machine(instance, [ 792.301669] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] for vif in network_info: [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] return self._sync_wrapper(fn, *args, **kwargs) [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self.wait() [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self[:] = self._gt.wait() [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] return self._exit_event.wait() [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 792.301987] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] current.throw(*self._exc) [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] result = function(*args, **kwargs) [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] return func(*args, **kwargs) [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] raise e [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] nwinfo = self.network_api.allocate_for_instance( [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] created_port_ids = self._update_ports_for_instance( [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] with excutils.save_and_reraise_exception(): [ 792.302309] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self.force_reraise() [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] raise self.value [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] updated_port = self._update_port( [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] _ensure_no_port_binding_failure(port) [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] raise exception.PortBindingFailed(port_id=port['id']) [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] nova.exception.PortBindingFailed: Binding failed for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6, please check neutron logs for more information. [ 792.302614] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] [ 792.302614] env[62183]: INFO nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Terminating instance [ 792.303771] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "refresh_cache-27ba0765-152b-4985-927d-99670818a0b1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.303942] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired lock "refresh_cache-27ba0765-152b-4985-927d-99670818a0b1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.304121] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.530333] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.556299] env[62183]: DEBUG oslo_concurrency.lockutils [req-9c4d1a91-21c8-4c40-8023-5ab257591ead req-7b6d07e1-d3a4-4e73-ad8c-9c8bae4ed763 service nova] Releasing lock "refresh_cache-bd5c642b-05e2-4c96-8e22-54b02fc93263" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.556508] env[62183]: DEBUG nova.compute.manager [req-9c4d1a91-21c8-4c40-8023-5ab257591ead req-7b6d07e1-d3a4-4e73-ad8c-9c8bae4ed763 service nova] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Received event network-vif-deleted-45b75c14-62b4-4d5e-be95-8e02039570c1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.607777] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.819128] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.887606] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.110476] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Releasing lock "refresh_cache-9278c981-042e-41a2-a4f8-d42a8e6fac39" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.110735] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 793.110941] env[62183]: DEBUG nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.111131] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.125623] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.210168] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d5cf81-ae93-4782-9d8a-602c4ba0d601 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.217744] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45444bf2-db6d-4d39-ba4c-073ece7cf638 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.247199] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a208017f-5269-4946-9191-57fbf3dae02a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.254190] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e6f898-1d2f-4212-be92-cb39c8d9b414 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.266941] env[62183]: DEBUG nova.compute.provider_tree [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.392082] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lock "refresh_cache-27ba0765-152b-4985-927d-99670818a0b1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.392586] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 793.392838] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.393167] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-873c5849-bd7e-47cd-bb44-582fc5b3d7ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.402573] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656e722c-2e04-4673-bd28-df484130bc08 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.424997] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 27ba0765-152b-4985-927d-99670818a0b1 could not be found. [ 793.425253] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 793.425435] env[62183]: INFO nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 793.425677] env[62183]: DEBUG oslo.service.loopingcall [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.425909] env[62183]: DEBUG nova.compute.manager [-] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.426015] env[62183]: DEBUG nova.network.neutron [-] [instance: 27ba0765-152b-4985-927d-99670818a0b1] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.439113] env[62183]: DEBUG nova.network.neutron [-] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.446173] env[62183]: DEBUG nova.compute.manager [req-321828a9-c4e1-4753-8980-f4decfd52b57 req-fed46b52-b128-4885-b2e3-b9df9595c008 service nova] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Received event network-changed-ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.446321] env[62183]: DEBUG nova.compute.manager [req-321828a9-c4e1-4753-8980-f4decfd52b57 req-fed46b52-b128-4885-b2e3-b9df9595c008 service nova] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Refreshing instance network info cache due to event network-changed-ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.446533] env[62183]: DEBUG oslo_concurrency.lockutils [req-321828a9-c4e1-4753-8980-f4decfd52b57 req-fed46b52-b128-4885-b2e3-b9df9595c008 service nova] Acquiring lock "refresh_cache-27ba0765-152b-4985-927d-99670818a0b1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.446661] env[62183]: DEBUG oslo_concurrency.lockutils [req-321828a9-c4e1-4753-8980-f4decfd52b57 req-fed46b52-b128-4885-b2e3-b9df9595c008 service nova] Acquired lock "refresh_cache-27ba0765-152b-4985-927d-99670818a0b1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.446815] env[62183]: DEBUG nova.network.neutron [req-321828a9-c4e1-4753-8980-f4decfd52b57 req-fed46b52-b128-4885-b2e3-b9df9595c008 service nova] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Refreshing network info cache for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 793.627836] env[62183]: DEBUG nova.network.neutron [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.769836] env[62183]: DEBUG nova.scheduler.client.report [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 793.941136] env[62183]: DEBUG nova.network.neutron [-] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.968126] env[62183]: DEBUG nova.network.neutron [req-321828a9-c4e1-4753-8980-f4decfd52b57 req-fed46b52-b128-4885-b2e3-b9df9595c008 service nova] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.035207] env[62183]: DEBUG nova.network.neutron [req-321828a9-c4e1-4753-8980-f4decfd52b57 req-fed46b52-b128-4885-b2e3-b9df9595c008 service nova] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.130266] env[62183]: INFO nova.compute.manager [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 9278c981-042e-41a2-a4f8-d42a8e6fac39] Took 1.02 seconds to deallocate network for instance. [ 794.275233] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.275635] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 794.278753] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.255s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.443270] env[62183]: INFO nova.compute.manager [-] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Took 1.02 seconds to deallocate network for instance. [ 794.445936] env[62183]: DEBUG nova.compute.claims [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 794.445936] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.538480] env[62183]: DEBUG oslo_concurrency.lockutils [req-321828a9-c4e1-4753-8980-f4decfd52b57 req-fed46b52-b128-4885-b2e3-b9df9595c008 service nova] Releasing lock "refresh_cache-27ba0765-152b-4985-927d-99670818a0b1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.538480] env[62183]: DEBUG nova.compute.manager [req-321828a9-c4e1-4753-8980-f4decfd52b57 req-fed46b52-b128-4885-b2e3-b9df9595c008 service nova] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Received event network-vif-deleted-ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.780029] env[62183]: DEBUG nova.compute.utils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.781464] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.781689] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 794.823644] env[62183]: DEBUG nova.policy [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca09642e8daf4d12ae1989d7369054c0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ed7ada746a14aa5b6ebe958dbb11339', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.989185] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb5553f-e1aa-4936-b003-c457fb80a35f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.996799] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58aa97b2-bd29-44c5-9a19-cf49dadba807 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.026749] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57752131-77c8-4f4b-b96f-3e8933b38c11 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.034207] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb7e1a8-73de-42b6-b527-e87032154117 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.047317] env[62183]: DEBUG nova.compute.provider_tree [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.202377] env[62183]: INFO nova.scheduler.client.report [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Deleted allocations for instance 9278c981-042e-41a2-a4f8-d42a8e6fac39 [ 795.285559] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 795.350805] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Successfully created port: 8f952cef-2a08-44ae-8f53-841a7833255f {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 795.550596] env[62183]: DEBUG nova.scheduler.client.report [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 795.712291] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0c8e8308-4035-42f8-a1f6-a58c07833b62 tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "9278c981-042e-41a2-a4f8-d42a8e6fac39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.166s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.791197] env[62183]: INFO nova.virt.block_device [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Booting with volume 7dd5c0dd-11ed-44e7-b84f-9afb7440e307 at /dev/sda [ 795.867452] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8b2c579-5d05-4ea5-82d8-7ca1c8a2ed4c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.876616] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3662a05a-21db-4c0d-a233-62bde6790d7d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.898143] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7db97add-09ee-468e-9902-53c1fa3dfc94 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.905862] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29a9414-7e8c-47d1-bea6-65abd5c48bcc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.928588] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b8a67a-a616-4c8c-b7ff-8570a0725b66 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.936116] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5896ba9-d4a4-47b7-b79e-7f1d3487a24c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.951683] env[62183]: DEBUG nova.virt.block_device [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Updating existing volume attachment record: 6191f093-629f-4001-a850-43040c1e4c1e {{(pid=62183) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 796.058930] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.780s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.059817] env[62183]: ERROR nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f5db769a-3499-47e8-9d21-7ada9e155260, please check neutron logs for more information. [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Traceback (most recent call last): [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self.driver.spawn(context, instance, image_meta, [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] vm_ref = self.build_virtual_machine(instance, [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 796.059817] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] for vif in network_info: [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] return self._sync_wrapper(fn, *args, **kwargs) [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self.wait() [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self[:] = self._gt.wait() [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] return self._exit_event.wait() [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] current.throw(*self._exc) [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.060348] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] result = function(*args, **kwargs) [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] return func(*args, **kwargs) [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] raise e [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] nwinfo = self.network_api.allocate_for_instance( [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] created_port_ids = self._update_ports_for_instance( [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] with excutils.save_and_reraise_exception(): [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] self.force_reraise() [ 796.060876] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.062450] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] raise self.value [ 796.062450] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 796.062450] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] updated_port = self._update_port( [ 796.062450] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.062450] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] _ensure_no_port_binding_failure(port) [ 796.062450] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.062450] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] raise exception.PortBindingFailed(port_id=port['id']) [ 796.062450] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] nova.exception.PortBindingFailed: Binding failed for port f5db769a-3499-47e8-9d21-7ada9e155260, please check neutron logs for more information. [ 796.062450] env[62183]: ERROR nova.compute.manager [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] [ 796.062450] env[62183]: DEBUG nova.compute.utils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Binding failed for port f5db769a-3499-47e8-9d21-7ada9e155260, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 796.063705] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.452s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.067036] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Build of instance 3b231f56-5127-4ff9-9031-06bfd59a57d0 was re-scheduled: Binding failed for port f5db769a-3499-47e8-9d21-7ada9e155260, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 796.067535] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 796.067822] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquiring lock "refresh_cache-3b231f56-5127-4ff9-9031-06bfd59a57d0" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.068038] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Acquired lock "refresh_cache-3b231f56-5127-4ff9-9031-06bfd59a57d0" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.068259] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.215124] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 796.349744] env[62183]: DEBUG nova.compute.manager [req-3d6dd7f7-0209-4b7d-9dec-5507e99642bb req-a7a55355-a0bd-496a-86f1-a3889d97837d service nova] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Received event network-changed-8f952cef-2a08-44ae-8f53-841a7833255f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.349951] env[62183]: DEBUG nova.compute.manager [req-3d6dd7f7-0209-4b7d-9dec-5507e99642bb req-a7a55355-a0bd-496a-86f1-a3889d97837d service nova] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Refreshing instance network info cache due to event network-changed-8f952cef-2a08-44ae-8f53-841a7833255f. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 796.350189] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d6dd7f7-0209-4b7d-9dec-5507e99642bb req-a7a55355-a0bd-496a-86f1-a3889d97837d service nova] Acquiring lock "refresh_cache-5a559b13-b1e0-414a-896a-e70648ed93d9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.350364] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d6dd7f7-0209-4b7d-9dec-5507e99642bb req-a7a55355-a0bd-496a-86f1-a3889d97837d service nova] Acquired lock "refresh_cache-5a559b13-b1e0-414a-896a-e70648ed93d9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.350490] env[62183]: DEBUG nova.network.neutron [req-3d6dd7f7-0209-4b7d-9dec-5507e99642bb req-a7a55355-a0bd-496a-86f1-a3889d97837d service nova] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Refreshing network info cache for port 8f952cef-2a08-44ae-8f53-841a7833255f {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.589293] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.685596] env[62183]: ERROR nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f952cef-2a08-44ae-8f53-841a7833255f, please check neutron logs for more information. [ 796.685596] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 796.685596] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.685596] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 796.685596] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 796.685596] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 796.685596] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 796.685596] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 796.685596] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.685596] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 796.685596] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.685596] env[62183]: ERROR nova.compute.manager raise self.value [ 796.685596] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 796.685596] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 796.685596] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.685596] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 796.686071] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.686071] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 796.686071] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f952cef-2a08-44ae-8f53-841a7833255f, please check neutron logs for more information. [ 796.686071] env[62183]: ERROR nova.compute.manager [ 796.686071] env[62183]: Traceback (most recent call last): [ 796.686071] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 796.686071] env[62183]: listener.cb(fileno) [ 796.686071] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.686071] env[62183]: result = function(*args, **kwargs) [ 796.686071] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 796.686071] env[62183]: return func(*args, **kwargs) [ 796.686071] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 796.686071] env[62183]: raise e [ 796.686071] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 796.686071] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 796.686071] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 796.686071] env[62183]: created_port_ids = self._update_ports_for_instance( [ 796.686071] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 796.686071] env[62183]: with excutils.save_and_reraise_exception(): [ 796.686071] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.686071] env[62183]: self.force_reraise() [ 796.686071] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.686071] env[62183]: raise self.value [ 796.686071] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 796.686071] env[62183]: updated_port = self._update_port( [ 796.686071] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.686071] env[62183]: _ensure_no_port_binding_failure(port) [ 796.686071] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.686071] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 796.686859] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 8f952cef-2a08-44ae-8f53-841a7833255f, please check neutron logs for more information. [ 796.686859] env[62183]: Removing descriptor: 14 [ 796.724721] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.732196] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.807259] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07f0ba1-174b-474f-b886-7f4e2c3e0fab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.815191] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f09ca3-2d08-4435-8e0e-bd6dc46f4ec7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.844277] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ff9176-a415-4903-84a9-79e80b8b7a22 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.851115] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56992af-db7e-4dd2-a0ba-34d114aa0fb7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.865491] env[62183]: DEBUG nova.compute.provider_tree [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.871691] env[62183]: DEBUG nova.network.neutron [req-3d6dd7f7-0209-4b7d-9dec-5507e99642bb req-a7a55355-a0bd-496a-86f1-a3889d97837d service nova] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.956948] env[62183]: DEBUG nova.network.neutron [req-3d6dd7f7-0209-4b7d-9dec-5507e99642bb req-a7a55355-a0bd-496a-86f1-a3889d97837d service nova] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.230068] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Releasing lock "refresh_cache-3b231f56-5127-4ff9-9031-06bfd59a57d0" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.230349] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 797.230543] env[62183]: DEBUG nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.230771] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 797.248776] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.369202] env[62183]: DEBUG nova.scheduler.client.report [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 797.459905] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d6dd7f7-0209-4b7d-9dec-5507e99642bb req-a7a55355-a0bd-496a-86f1-a3889d97837d service nova] Releasing lock "refresh_cache-5a559b13-b1e0-414a-896a-e70648ed93d9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.751677] env[62183]: DEBUG nova.network.neutron [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.877097] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.813s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.877097] env[62183]: ERROR nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c7364122-69c2-4c96-8086-736513d9265e, please check neutron logs for more information. [ 797.877097] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Traceback (most recent call last): [ 797.877097] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 797.877097] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self.driver.spawn(context, instance, image_meta, [ 797.877097] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 797.877097] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 797.877097] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 797.877097] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] vm_ref = self.build_virtual_machine(instance, [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] for vif in network_info: [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] return self._sync_wrapper(fn, *args, **kwargs) [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self.wait() [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self[:] = self._gt.wait() [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] return self._exit_event.wait() [ 797.877405] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] result = hub.switch() [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] return self.greenlet.switch() [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] result = function(*args, **kwargs) [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] return func(*args, **kwargs) [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] raise e [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] nwinfo = self.network_api.allocate_for_instance( [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 797.877922] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] created_port_ids = self._update_ports_for_instance( [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] with excutils.save_and_reraise_exception(): [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] self.force_reraise() [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] raise self.value [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] updated_port = self._update_port( [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] _ensure_no_port_binding_failure(port) [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 797.878427] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] raise exception.PortBindingFailed(port_id=port['id']) [ 797.878860] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] nova.exception.PortBindingFailed: Binding failed for port c7364122-69c2-4c96-8086-736513d9265e, please check neutron logs for more information. [ 797.878860] env[62183]: ERROR nova.compute.manager [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] [ 797.878860] env[62183]: DEBUG nova.compute.utils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Binding failed for port c7364122-69c2-4c96-8086-736513d9265e, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 797.878860] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.425s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.881649] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Build of instance 3fef2148-244d-4575-8e3a-93bf70c7dd7c was re-scheduled: Binding failed for port c7364122-69c2-4c96-8086-736513d9265e, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 797.882073] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 797.882296] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Acquiring lock "refresh_cache-3fef2148-244d-4575-8e3a-93bf70c7dd7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.882442] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Acquired lock "refresh_cache-3fef2148-244d-4575-8e3a-93bf70c7dd7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.882597] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 798.101129] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 798.101686] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.101896] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.102064] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.102248] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.102392] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.102536] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.102734] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.102920] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.103100] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.103263] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.103433] env[62183]: DEBUG nova.virt.hardware [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.104305] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f6d765-ed77-4130-9ae3-1e41e9982ad0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.112551] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664d7a6e-abb4-4547-9d96-6c552ffbce38 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.125974] env[62183]: ERROR nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f952cef-2a08-44ae-8f53-841a7833255f, please check neutron logs for more information. [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Traceback (most recent call last): [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] yield resources [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self.driver.spawn(context, instance, image_meta, [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] vm_ref = self.build_virtual_machine(instance, [ 798.125974] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] for vif in network_info: [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] return self._sync_wrapper(fn, *args, **kwargs) [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self.wait() [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self[:] = self._gt.wait() [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] return self._exit_event.wait() [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 798.126348] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] current.throw(*self._exc) [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] result = function(*args, **kwargs) [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] return func(*args, **kwargs) [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] raise e [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] nwinfo = self.network_api.allocate_for_instance( [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] created_port_ids = self._update_ports_for_instance( [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] with excutils.save_and_reraise_exception(): [ 798.126700] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self.force_reraise() [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] raise self.value [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] updated_port = self._update_port( [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] _ensure_no_port_binding_failure(port) [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] raise exception.PortBindingFailed(port_id=port['id']) [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] nova.exception.PortBindingFailed: Binding failed for port 8f952cef-2a08-44ae-8f53-841a7833255f, please check neutron logs for more information. [ 798.127070] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] [ 798.127070] env[62183]: INFO nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Terminating instance [ 798.129904] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Acquiring lock "refresh_cache-5a559b13-b1e0-414a-896a-e70648ed93d9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.130084] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Acquired lock "refresh_cache-5a559b13-b1e0-414a-896a-e70648ed93d9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.130255] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 798.254280] env[62183]: INFO nova.compute.manager [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] [instance: 3b231f56-5127-4ff9-9031-06bfd59a57d0] Took 1.02 seconds to deallocate network for instance. [ 798.374768] env[62183]: DEBUG nova.compute.manager [req-7b53b61a-ac34-419b-82ff-7b0614d111cc req-55d2cc60-a8a7-416b-9076-5cc0ab3380c8 service nova] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Received event network-vif-deleted-8f952cef-2a08-44ae-8f53-841a7833255f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.402690] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.532019] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.572316] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6553da1c-c704-41aa-b983-adfd23a21da9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.579762] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188860c9-9852-493d-9aa5-7cbaeeca0b42 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.610065] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79ee390-8174-493a-9e60-e00f8a6b5624 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.617169] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55afd41f-377e-4653-941a-3b925db946e8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.629863] env[62183]: DEBUG nova.compute.provider_tree [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.646259] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.726358] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.036765] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Releasing lock "refresh_cache-3fef2148-244d-4575-8e3a-93bf70c7dd7c" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.037049] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 799.037251] env[62183]: DEBUG nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.037423] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 799.051632] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.133176] env[62183]: DEBUG nova.scheduler.client.report [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 799.229082] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Releasing lock "refresh_cache-5a559b13-b1e0-414a-896a-e70648ed93d9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.229682] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 799.230011] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c420d747-e77c-4b9c-a614-e6754f5b298b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.239209] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95986f41-1a37-476f-93ad-9c27bec88f0b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.263939] env[62183]: WARNING nova.virt.vmwareapi.driver [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 5a559b13-b1e0-414a-896a-e70648ed93d9 could not be found. [ 799.264225] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 799.264601] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb43d75c-4a75-4ee7-9484-8b3cbe61e0a0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.271922] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5fb284-06ed-4134-ad83-3d5494894a9a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.289009] env[62183]: INFO nova.scheduler.client.report [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Deleted allocations for instance 3b231f56-5127-4ff9-9031-06bfd59a57d0 [ 799.298012] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5a559b13-b1e0-414a-896a-e70648ed93d9 could not be found. [ 799.298221] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 799.298399] env[62183]: INFO nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Took 0.07 seconds to destroy the instance on the hypervisor. [ 799.298787] env[62183]: DEBUG oslo.service.loopingcall [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.299624] env[62183]: DEBUG nova.compute.manager [-] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.299723] env[62183]: DEBUG nova.network.neutron [-] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 799.314591] env[62183]: DEBUG nova.network.neutron [-] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.554481] env[62183]: DEBUG nova.network.neutron [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.638316] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.759s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.638933] env[62183]: ERROR nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 54fe49a1-3468-438a-86c6-129f68209fa1, please check neutron logs for more information. [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Traceback (most recent call last): [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self.driver.spawn(context, instance, image_meta, [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self._vmops.spawn(context, instance, image_meta, injected_files, [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] vm_ref = self.build_virtual_machine(instance, [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] vif_infos = vmwarevif.get_vif_info(self._session, [ 799.638933] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] for vif in network_info: [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] return self._sync_wrapper(fn, *args, **kwargs) [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self.wait() [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self[:] = self._gt.wait() [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] return self._exit_event.wait() [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] current.throw(*self._exc) [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 799.639402] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] result = function(*args, **kwargs) [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] return func(*args, **kwargs) [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] raise e [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] nwinfo = self.network_api.allocate_for_instance( [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] created_port_ids = self._update_ports_for_instance( [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] with excutils.save_and_reraise_exception(): [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] self.force_reraise() [ 799.639719] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 799.640087] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] raise self.value [ 799.640087] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 799.640087] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] updated_port = self._update_port( [ 799.640087] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 799.640087] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] _ensure_no_port_binding_failure(port) [ 799.640087] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 799.640087] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] raise exception.PortBindingFailed(port_id=port['id']) [ 799.640087] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] nova.exception.PortBindingFailed: Binding failed for port 54fe49a1-3468-438a-86c6-129f68209fa1, please check neutron logs for more information. [ 799.640087] env[62183]: ERROR nova.compute.manager [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] [ 799.640087] env[62183]: DEBUG nova.compute.utils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Binding failed for port 54fe49a1-3468-438a-86c6-129f68209fa1, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 799.640913] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.815s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.644195] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Build of instance 48909506-5abd-4dfd-9aac-f3c1ef0f8250 was re-scheduled: Binding failed for port 54fe49a1-3468-438a-86c6-129f68209fa1, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 799.644615] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 799.644861] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquiring lock "refresh_cache-48909506-5abd-4dfd-9aac-f3c1ef0f8250" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.645031] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Acquired lock "refresh_cache-48909506-5abd-4dfd-9aac-f3c1ef0f8250" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.645197] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.803164] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bfaa327d-5b33-4c9f-90c3-5bd876726f3b tempest-ServerRescueNegativeTestJSON-650720359 tempest-ServerRescueNegativeTestJSON-650720359-project-member] Lock "3b231f56-5127-4ff9-9031-06bfd59a57d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.785s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.817072] env[62183]: DEBUG nova.network.neutron [-] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.057137] env[62183]: INFO nova.compute.manager [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] [instance: 3fef2148-244d-4575-8e3a-93bf70c7dd7c] Took 1.02 seconds to deallocate network for instance. [ 800.166677] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.218965] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.304780] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.319060] env[62183]: INFO nova.compute.manager [-] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Took 1.02 seconds to deallocate network for instance. [ 800.322290] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaafab3d-b1a8-42e5-a27f-eb27e5cd342b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.329422] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f6d16c-b8b7-4dc0-8086-a1f7792e63ff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.359789] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd662e0-446d-42d4-8a10-ac9b3ccbe7c9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.367017] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc82b854-46df-42b6-8fe9-5bd85c564492 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.380201] env[62183]: DEBUG nova.compute.provider_tree [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.722051] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Releasing lock "refresh_cache-48909506-5abd-4dfd-9aac-f3c1ef0f8250" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.722335] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 800.722503] env[62183]: DEBUG nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 800.722673] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 800.740021] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.833581] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.874182] env[62183]: INFO nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Took 0.55 seconds to detach 1 volumes for instance. [ 800.876370] env[62183]: DEBUG nova.compute.claims [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 800.876604] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.882954] env[62183]: DEBUG nova.scheduler.client.report [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 801.094358] env[62183]: INFO nova.scheduler.client.report [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Deleted allocations for instance 3fef2148-244d-4575-8e3a-93bf70c7dd7c [ 801.242510] env[62183]: DEBUG nova.network.neutron [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.387226] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.746s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.387860] env[62183]: ERROR nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2358598c-6738-475e-b6bf-5be07a6070a6, please check neutron logs for more information. [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Traceback (most recent call last): [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self.driver.spawn(context, instance, image_meta, [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] vm_ref = self.build_virtual_machine(instance, [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] vif_infos = vmwarevif.get_vif_info(self._session, [ 801.387860] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] for vif in network_info: [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] return self._sync_wrapper(fn, *args, **kwargs) [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self.wait() [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self[:] = self._gt.wait() [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] return self._exit_event.wait() [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] current.throw(*self._exc) [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 801.388231] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] result = function(*args, **kwargs) [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] return func(*args, **kwargs) [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] raise e [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] nwinfo = self.network_api.allocate_for_instance( [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] created_port_ids = self._update_ports_for_instance( [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] with excutils.save_and_reraise_exception(): [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] self.force_reraise() [ 801.388577] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.388918] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] raise self.value [ 801.388918] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 801.388918] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] updated_port = self._update_port( [ 801.388918] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.388918] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] _ensure_no_port_binding_failure(port) [ 801.388918] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.388918] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] raise exception.PortBindingFailed(port_id=port['id']) [ 801.388918] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] nova.exception.PortBindingFailed: Binding failed for port 2358598c-6738-475e-b6bf-5be07a6070a6, please check neutron logs for more information. [ 801.388918] env[62183]: ERROR nova.compute.manager [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] [ 801.388918] env[62183]: DEBUG nova.compute.utils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Binding failed for port 2358598c-6738-475e-b6bf-5be07a6070a6, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 801.392086] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.326s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.392128] env[62183]: INFO nova.compute.claims [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.394652] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Build of instance 90be0169-9414-431b-9aa3-82d89bea4bb2 was re-scheduled: Binding failed for port 2358598c-6738-475e-b6bf-5be07a6070a6, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 801.395088] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 801.395291] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Acquiring lock "refresh_cache-90be0169-9414-431b-9aa3-82d89bea4bb2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.395430] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Acquired lock "refresh_cache-90be0169-9414-431b-9aa3-82d89bea4bb2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.395580] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.604579] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc0e7bd5-c5ce-4a2d-a17b-e9a736bfb1ec tempest-ServerPasswordTestJSON-897327113 tempest-ServerPasswordTestJSON-897327113-project-member] Lock "3fef2148-244d-4575-8e3a-93bf70c7dd7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.127s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.745133] env[62183]: INFO nova.compute.manager [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] [instance: 48909506-5abd-4dfd-9aac-f3c1ef0f8250] Took 1.02 seconds to deallocate network for instance. [ 801.920919] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.079755] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.106916] env[62183]: DEBUG nova.compute.manager [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 802.583065] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Releasing lock "refresh_cache-90be0169-9414-431b-9aa3-82d89bea4bb2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.583562] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 802.583759] env[62183]: DEBUG nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.583923] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.606192] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.620234] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf27ee09-aa48-4f65-a189-d0161c2dea04 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.625994] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.629016] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166ceb8b-86ef-4e72-b72c-403fbc42c422 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.660070] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13028f90-8147-4965-b70a-f7aa12c8a2c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.667743] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194adc94-0a9d-4597-a3b3-e1b19285bb5c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.681729] env[62183]: DEBUG nova.compute.provider_tree [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.779184] env[62183]: INFO nova.scheduler.client.report [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Deleted allocations for instance 48909506-5abd-4dfd-9aac-f3c1ef0f8250 [ 803.110890] env[62183]: DEBUG nova.network.neutron [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.185436] env[62183]: DEBUG nova.scheduler.client.report [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 803.286453] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2dcd0856-75e2-423e-94fe-9c1722dd7e7f tempest-DeleteServersAdminTestJSON-1711182313 tempest-DeleteServersAdminTestJSON-1711182313-project-member] Lock "48909506-5abd-4dfd-9aac-f3c1ef0f8250" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.364s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.617581] env[62183]: INFO nova.compute.manager [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] [instance: 90be0169-9414-431b-9aa3-82d89bea4bb2] Took 1.03 seconds to deallocate network for instance. [ 803.692109] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.300s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.692109] env[62183]: DEBUG nova.compute.manager [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.694913] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.997s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.792127] env[62183]: DEBUG nova.compute.manager [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 804.202187] env[62183]: DEBUG nova.compute.utils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.203859] env[62183]: DEBUG nova.compute.manager [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Not allocating networking since 'none' was specified. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 804.313094] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.388196] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ccce7f-9924-477f-9741-bc96599d88cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.396097] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f21ed5-dab4-4234-a8d7-77c0463ddb49 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.425754] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee862f7-006f-4dac-8eb2-9b1b11fc8af2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.432800] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770aa158-b742-4a7f-80c9-2634da2ef10c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.447293] env[62183]: DEBUG nova.compute.provider_tree [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.645969] env[62183]: INFO nova.scheduler.client.report [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Deleted allocations for instance 90be0169-9414-431b-9aa3-82d89bea4bb2 [ 804.705746] env[62183]: DEBUG nova.compute.manager [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.950655] env[62183]: DEBUG nova.scheduler.client.report [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 805.155269] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e048626f-44d5-4b37-856d-7d23994a496f tempest-ServerActionsTestOtherA-113356685 tempest-ServerActionsTestOtherA-113356685-project-member] Lock "90be0169-9414-431b-9aa3-82d89bea4bb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.199s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.371491] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "4fa0beb2-3f36-48f4-996c-1da0bb891f93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.371770] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "4fa0beb2-3f36-48f4-996c-1da0bb891f93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.455444] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.760s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.456095] env[62183]: ERROR nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 39ddd09f-b5f7-419f-8415-69db4af4d169, please check neutron logs for more information. [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Traceback (most recent call last): [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self.driver.spawn(context, instance, image_meta, [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self._vmops.spawn(context, instance, image_meta, injected_files, [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] vm_ref = self.build_virtual_machine(instance, [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] vif_infos = vmwarevif.get_vif_info(self._session, [ 805.456095] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] for vif in network_info: [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] return self._sync_wrapper(fn, *args, **kwargs) [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self.wait() [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self[:] = self._gt.wait() [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] return self._exit_event.wait() [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] current.throw(*self._exc) [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 805.456475] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] result = function(*args, **kwargs) [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] return func(*args, **kwargs) [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] raise e [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] nwinfo = self.network_api.allocate_for_instance( [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] created_port_ids = self._update_ports_for_instance( [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] with excutils.save_and_reraise_exception(): [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] self.force_reraise() [ 805.456809] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 805.457164] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] raise self.value [ 805.457164] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 805.457164] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] updated_port = self._update_port( [ 805.457164] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 805.457164] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] _ensure_no_port_binding_failure(port) [ 805.457164] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 805.457164] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] raise exception.PortBindingFailed(port_id=port['id']) [ 805.457164] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] nova.exception.PortBindingFailed: Binding failed for port 39ddd09f-b5f7-419f-8415-69db4af4d169, please check neutron logs for more information. [ 805.457164] env[62183]: ERROR nova.compute.manager [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] [ 805.457164] env[62183]: DEBUG nova.compute.utils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Binding failed for port 39ddd09f-b5f7-419f-8415-69db4af4d169, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 805.457981] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.372s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.459794] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Build of instance fe04dc9a-9fda-41e8-b62e-78dc4c026968 was re-scheduled: Binding failed for port 39ddd09f-b5f7-419f-8415-69db4af4d169, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 805.460110] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 805.460331] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Acquiring lock "refresh_cache-fe04dc9a-9fda-41e8-b62e-78dc4c026968" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.460474] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Acquired lock "refresh_cache-fe04dc9a-9fda-41e8-b62e-78dc4c026968" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.460629] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.657862] env[62183]: DEBUG nova.compute.manager [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.715269] env[62183]: DEBUG nova.compute.manager [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.739797] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.740059] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.740222] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.740403] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.740549] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.740693] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.740896] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.741067] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.741238] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.741399] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.741567] env[62183]: DEBUG nova.virt.hardware [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.742437] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66be663d-d622-468b-b792-b59fdd8f0d9f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.750301] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784ec298-4c6b-42f6-bacc-df872db08895 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.765060] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.769846] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Creating folder: Project (0d97dcec7e39447f9166954bc4e6b365). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.770140] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-098fac25-5d15-40b6-8a08-1376a3ebac71 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.781239] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Created folder: Project (0d97dcec7e39447f9166954bc4e6b365) in parent group-v294392. [ 805.781427] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Creating folder: Instances. Parent ref: group-v294417. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.781651] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43deb9bd-008d-4dd5-b340-ed8bf685e5af {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.790236] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Created folder: Instances in parent group-v294417. [ 805.790464] env[62183]: DEBUG oslo.service.loopingcall [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.790677] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07899178-b53c-4b0e-877d-11774286212f] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.790849] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-742326d2-5492-4b56-87bb-41990c33af32 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.808239] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.808239] env[62183]: value = "task-1386961" [ 805.808239] env[62183]: _type = "Task" [ 805.808239] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.815642] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386961, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.988271] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.118294] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.179534] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.319354] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386961, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.619839] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Releasing lock "refresh_cache-fe04dc9a-9fda-41e8-b62e-78dc4c026968" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.620132] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 806.620323] env[62183]: DEBUG nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.620491] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 806.635411] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.819093] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386961, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.993300] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance fe04dc9a-9fda-41e8-b62e-78dc4c026968 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 806.993300] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance bd5c642b-05e2-4c96-8e22-54b02fc93263 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.993300] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 27ba0765-152b-4985-927d-99670818a0b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.993300] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 5a559b13-b1e0-414a-896a-e70648ed93d9 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.993531] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 07899178-b53c-4b0e-877d-11774286212f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 807.138065] env[62183]: DEBUG nova.network.neutron [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.322700] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386961, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.498841] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 0dd90b88-ed92-4331-98c5-c7481ac1ae15 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 807.643448] env[62183]: INFO nova.compute.manager [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] [instance: fe04dc9a-9fda-41e8-b62e-78dc4c026968] Took 1.02 seconds to deallocate network for instance. [ 807.825665] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386961, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.005365] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 808.321428] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386961, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.508518] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 7465df88-c68c-49d3-9a91-6fff0d06957b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 808.682682] env[62183]: INFO nova.scheduler.client.report [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Deleted allocations for instance fe04dc9a-9fda-41e8-b62e-78dc4c026968 [ 808.823641] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386961, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.011777] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance ac555ffc-ce4e-4650-97fd-c26a3246fe4b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 809.196117] env[62183]: DEBUG oslo_concurrency.lockutils [None req-387b324c-c56d-4a92-9555-b7f04839a20a tempest-ServerMetadataTestJSON-1478141660 tempest-ServerMetadataTestJSON-1478141660-project-member] Lock "fe04dc9a-9fda-41e8-b62e-78dc4c026968" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.954s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.325913] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386961, 'name': CreateVM_Task, 'duration_secs': 3.255074} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.326128] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07899178-b53c-4b0e-877d-11774286212f] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.326576] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.326840] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.327098] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.327659] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-115200c8-bf18-425e-af36-9661df5035ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.332662] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 809.332662] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52d8a68b-072b-4919-8740-6ce50bbdcd4e" [ 809.332662] env[62183]: _type = "Task" [ 809.332662] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.341771] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d8a68b-072b-4919-8740-6ce50bbdcd4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.515400] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 68e791b4-61db-4b6b-a30a-ccf5d6657643 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 809.698692] env[62183]: DEBUG nova.compute.manager [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.846870] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d8a68b-072b-4919-8740-6ce50bbdcd4e, 'name': SearchDatastore_Task, 'duration_secs': 0.012851} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.847735] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.847984] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.848273] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.848427] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.848604] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.848877] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f96cb9ab-827b-4b14-9448-909e1e6272d7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.857118] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.857296] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.857997] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e82cecd5-0f9f-449d-b119-1bbab8680271 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.862937] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 809.862937] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52997437-e053-6144-a733-ccd8e2b57556" [ 809.862937] env[62183]: _type = "Task" [ 809.862937] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.870324] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52997437-e053-6144-a733-ccd8e2b57556, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.019019] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 39683931-d0ff-4a5c-a4a2-792230ab0e3d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 810.224956] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.380927] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52997437-e053-6144-a733-ccd8e2b57556, 'name': SearchDatastore_Task, 'duration_secs': 0.008348} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.381947] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6f83ce8-9576-44c2-8b81-d106f8e60708 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.388226] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 810.388226] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52487350-b45f-8506-3485-06e068fe45f7" [ 810.388226] env[62183]: _type = "Task" [ 810.388226] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.396479] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52487350-b45f-8506-3485-06e068fe45f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.525560] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 850d40d8-565a-49a2-a27f-3de2a8dc7e30 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 810.902932] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52487350-b45f-8506-3485-06e068fe45f7, 'name': SearchDatastore_Task, 'duration_secs': 0.009032} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.904469] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.905622] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 07899178-b53c-4b0e-877d-11774286212f/07899178-b53c-4b0e-877d-11774286212f.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.905622] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f17e12a8-c8b1-480b-89bd-22a5fd4312fb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.913048] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 810.913048] env[62183]: value = "task-1386962" [ 810.913048] env[62183]: _type = "Task" [ 810.913048] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.922981] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386962, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.031714] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 811.427406] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386962, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.534925] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 4fa0beb2-3f36-48f4-996c-1da0bb891f93 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 811.534925] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 811.534925] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 811.631694] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquiring lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.632493] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.769603] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c851e1-c529-40d5-b02d-c1187d9140f0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.776897] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635216aa-7294-4770-8414-122bacf5a564 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.806798] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc58ee5-6dee-4902-a56d-79410240c6e1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.814784] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba5f6e5-e4f8-4c2d-8ee0-96dc74d350d4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.828272] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.924747] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386962, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.788452} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.925030] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 07899178-b53c-4b0e-877d-11774286212f/07899178-b53c-4b0e-877d-11774286212f.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.925228] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.925476] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-98481109-7241-4008-bf40-20cdf4f499a9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.932078] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 811.932078] env[62183]: value = "task-1386963" [ 811.932078] env[62183]: _type = "Task" [ 811.932078] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.940506] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386963, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.331522] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 812.443348] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386963, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070948} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.443635] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.444443] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3966e589-7112-418d-9614-39feab6fbe3c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.464798] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 07899178-b53c-4b0e-877d-11774286212f/07899178-b53c-4b0e-877d-11774286212f.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.465110] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7810dd84-44f4-40b4-ae5b-b3fa05c8931e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.491578] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 812.491578] env[62183]: value = "task-1386964" [ 812.491578] env[62183]: _type = "Task" [ 812.491578] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.502141] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386964, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.837550] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 812.837806] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.380s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.838093] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.750s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.002149] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386964, 'name': ReconfigVM_Task, 'duration_secs': 0.249655} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.002149] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 07899178-b53c-4b0e-877d-11774286212f/07899178-b53c-4b0e-877d-11774286212f.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.002261] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-599151a4-d0bf-4fcd-94f4-72b4f9248c59 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.008021] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 813.008021] env[62183]: value = "task-1386965" [ 813.008021] env[62183]: _type = "Task" [ 813.008021] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.016440] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386965, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.524018] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386965, 'name': Rename_Task, 'duration_secs': 0.133048} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.525470] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.525943] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33c93822-0a73-47d8-9f5c-0a8a18ec1150 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.533695] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 813.533695] env[62183]: value = "task-1386966" [ 813.533695] env[62183]: _type = "Task" [ 813.533695] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.541980] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386966, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.584198] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4c500b-49c3-4612-99b0-f37519487079 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.589651] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7021c2-7a61-4792-b365-4f36e4f5d841 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.624176] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893c5f74-6974-42cf-8e6c-39955102bc95 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.631451] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74424bf-3645-497a-bec0-afbd82c34a3e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.645616] env[62183]: DEBUG nova.compute.provider_tree [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.654038] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "a948464b-63aa-4bc8-9885-228049e96d37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.654308] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.043161] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386966, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.148529] env[62183]: DEBUG nova.scheduler.client.report [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 814.543721] env[62183]: DEBUG oslo_vmware.api [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386966, 'name': PowerOnVM_Task, 'duration_secs': 0.535192} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.544059] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.544228] env[62183]: INFO nova.compute.manager [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Took 8.83 seconds to spawn the instance on the hypervisor. [ 814.544414] env[62183]: DEBUG nova.compute.manager [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.545168] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630bae33-23ab-4e03-873a-f5f446d7d8a1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.654266] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.816s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.654897] env[62183]: ERROR nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 45b75c14-62b4-4d5e-be95-8e02039570c1, please check neutron logs for more information. [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Traceback (most recent call last): [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self.driver.spawn(context, instance, image_meta, [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] vm_ref = self.build_virtual_machine(instance, [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.654897] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] for vif in network_info: [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] return self._sync_wrapper(fn, *args, **kwargs) [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self.wait() [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self[:] = self._gt.wait() [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] return self._exit_event.wait() [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] current.throw(*self._exc) [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.655221] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] result = function(*args, **kwargs) [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] return func(*args, **kwargs) [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] raise e [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] nwinfo = self.network_api.allocate_for_instance( [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] created_port_ids = self._update_ports_for_instance( [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] with excutils.save_and_reraise_exception(): [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] self.force_reraise() [ 814.655580] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.655920] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] raise self.value [ 814.655920] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.655920] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] updated_port = self._update_port( [ 814.655920] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.655920] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] _ensure_no_port_binding_failure(port) [ 814.655920] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.655920] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] raise exception.PortBindingFailed(port_id=port['id']) [ 814.655920] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] nova.exception.PortBindingFailed: Binding failed for port 45b75c14-62b4-4d5e-be95-8e02039570c1, please check neutron logs for more information. [ 814.655920] env[62183]: ERROR nova.compute.manager [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] [ 814.655920] env[62183]: DEBUG nova.compute.utils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Binding failed for port 45b75c14-62b4-4d5e-be95-8e02039570c1, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 814.656762] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.211s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.659540] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Build of instance bd5c642b-05e2-4c96-8e22-54b02fc93263 was re-scheduled: Binding failed for port 45b75c14-62b4-4d5e-be95-8e02039570c1, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 814.659934] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 814.660173] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-bd5c642b-05e2-4c96-8e22-54b02fc93263" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.660319] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-bd5c642b-05e2-4c96-8e22-54b02fc93263" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.660476] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.060925] env[62183]: INFO nova.compute.manager [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Took 29.01 seconds to build instance. [ 815.182499] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.266918] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.355744] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3da783-2fc9-4e6a-93df-f3883ff9cae2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.363278] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec575c5-149b-4f3d-a6ab-16c2dffbd9aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.391436] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a5dc34-2b03-401c-b00a-9611b798e12f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.398120] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0ab506-7bfe-4d18-ad8a-dbb422855647 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.410463] env[62183]: DEBUG nova.compute.provider_tree [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.562350] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d078afcd-04c0-407b-b258-4e57c6c87c68 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "07899178-b53c-4b0e-877d-11774286212f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.685s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.770468] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-bd5c642b-05e2-4c96-8e22-54b02fc93263" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.770729] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 815.770934] env[62183]: DEBUG nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.771107] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.786194] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.913841] env[62183]: DEBUG nova.scheduler.client.report [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 816.064800] env[62183]: DEBUG nova.compute.manager [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.289446] env[62183]: DEBUG nova.network.neutron [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.336304] env[62183]: INFO nova.compute.manager [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Rebuilding instance [ 816.374529] env[62183]: DEBUG nova.compute.manager [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.375452] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f0467f-1666-4526-b1e9-d9a30731c6d1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.418291] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.761s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.418967] env[62183]: ERROR nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6, please check neutron logs for more information. [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] Traceback (most recent call last): [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self.driver.spawn(context, instance, image_meta, [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] vm_ref = self.build_virtual_machine(instance, [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] vif_infos = vmwarevif.get_vif_info(self._session, [ 816.418967] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] for vif in network_info: [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] return self._sync_wrapper(fn, *args, **kwargs) [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self.wait() [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self[:] = self._gt.wait() [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] return self._exit_event.wait() [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] current.throw(*self._exc) [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 816.419277] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] result = function(*args, **kwargs) [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] return func(*args, **kwargs) [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] raise e [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] nwinfo = self.network_api.allocate_for_instance( [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] created_port_ids = self._update_ports_for_instance( [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] with excutils.save_and_reraise_exception(): [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] self.force_reraise() [ 816.419741] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.420101] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] raise self.value [ 816.420101] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 816.420101] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] updated_port = self._update_port( [ 816.420101] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.420101] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] _ensure_no_port_binding_failure(port) [ 816.420101] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.420101] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] raise exception.PortBindingFailed(port_id=port['id']) [ 816.420101] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] nova.exception.PortBindingFailed: Binding failed for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6, please check neutron logs for more information. [ 816.420101] env[62183]: ERROR nova.compute.manager [instance: 27ba0765-152b-4985-927d-99670818a0b1] [ 816.420101] env[62183]: DEBUG nova.compute.utils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Binding failed for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 816.420948] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.689s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.422734] env[62183]: INFO nova.compute.claims [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.425923] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Build of instance 27ba0765-152b-4985-927d-99670818a0b1 was re-scheduled: Binding failed for port ba3dc34f-4c1a-44f1-9d71-699c1cf6a4b6, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 816.426446] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 816.426712] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "refresh_cache-27ba0765-152b-4985-927d-99670818a0b1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.426919] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired lock "refresh_cache-27ba0765-152b-4985-927d-99670818a0b1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.427129] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.583147] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.792521] env[62183]: INFO nova.compute.manager [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: bd5c642b-05e2-4c96-8e22-54b02fc93263] Took 1.02 seconds to deallocate network for instance. [ 816.890270] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 816.890702] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0ab248f-e9cd-4002-92e1-f1356aa6b664 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.898752] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 816.898752] env[62183]: value = "task-1386967" [ 816.898752] env[62183]: _type = "Task" [ 816.898752] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.908520] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386967, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.954159] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.045172] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.408465] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386967, 'name': PowerOffVM_Task, 'duration_secs': 0.113227} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.408739] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 817.408952] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 817.409725] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b012104-582c-4443-bcc7-2ef91d0bb35c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.416290] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 817.416530] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff2704e4-2d67-4feb-80ca-40fba357da27 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.440746] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 817.440952] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 817.441250] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Deleting the datastore file [datastore1] 07899178-b53c-4b0e-877d-11774286212f {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 817.441503] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7498591-bacd-44e1-ac1b-33b496677487 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.447288] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 817.447288] env[62183]: value = "task-1386969" [ 817.447288] env[62183]: _type = "Task" [ 817.447288] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.454871] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386969, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.547649] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lock "refresh_cache-27ba0765-152b-4985-927d-99670818a0b1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.547891] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 817.548591] env[62183]: DEBUG nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.548591] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 817.567650] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.612160] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef77a858-0a01-4ccf-8b31-d66ebdd29df8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.620459] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93da369f-09cd-48bc-8e86-602448b1da6f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.649071] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e645366-2f56-43ac-9f3a-a1154547df15 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.655734] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c77d7ad-8137-402c-bfe3-5f7c3fe173c0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.668276] env[62183]: DEBUG nova.compute.provider_tree [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.824902] env[62183]: INFO nova.scheduler.client.report [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted allocations for instance bd5c642b-05e2-4c96-8e22-54b02fc93263 [ 817.957180] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386969, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120916} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.957502] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.958921] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 817.958921] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 818.069986] env[62183]: DEBUG nova.network.neutron [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.171401] env[62183]: DEBUG nova.scheduler.client.report [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 818.332777] env[62183]: DEBUG oslo_concurrency.lockutils [None req-72fad47e-646e-4a2f-ad0c-c9726f23267e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "bd5c642b-05e2-4c96-8e22-54b02fc93263" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.992s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.574017] env[62183]: INFO nova.compute.manager [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: 27ba0765-152b-4985-927d-99670818a0b1] Took 1.02 seconds to deallocate network for instance. [ 818.676416] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.255s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.676953] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.679393] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.846s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.680763] env[62183]: INFO nova.compute.claims [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.836127] env[62183]: DEBUG nova.compute.manager [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.987181] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.987466] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.987626] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.987805] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.987946] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.988115] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.988329] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.988485] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.988648] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.988806] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.988969] env[62183]: DEBUG nova.virt.hardware [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.989813] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe0cc1a-68aa-4586-9fa8-879fd6f67e27 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.997829] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977d3665-84e4-4d32-855a-6d9bd3680022 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.011303] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.016871] env[62183]: DEBUG oslo.service.loopingcall [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.017135] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07899178-b53c-4b0e-877d-11774286212f] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 819.017494] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-85023d1e-fb38-4b02-bd6f-9a29bdb63fdc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.033864] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.033864] env[62183]: value = "task-1386970" [ 819.033864] env[62183]: _type = "Task" [ 819.033864] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.042017] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386970, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.186020] env[62183]: DEBUG nova.compute.utils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.189073] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 819.189073] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 819.228524] env[62183]: DEBUG nova.policy [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03a2b9f0e7eb4ac7be6d7708e3b847cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df2801b557804752a6fca3b5c2ff4d88', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.355456] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.544256] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386970, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.565075] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Successfully created port: abf57cfb-a5e5-4710-8eeb-bec026db61f5 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.567295] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.567651] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.605195] env[62183]: INFO nova.scheduler.client.report [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Deleted allocations for instance 27ba0765-152b-4985-927d-99670818a0b1 [ 819.690386] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.872993] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6161ac-cbcd-424c-b8b7-8e7ede9e7fba {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.880775] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20bf239-bd20-4da5-9046-cd687665d721 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.919106] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3410da05-ec5f-4a1a-a146-5cfe9bc7a5c7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.927012] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969bf6b9-222b-4629-b0e6-d9c428c0ed59 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.940588] env[62183]: DEBUG nova.compute.provider_tree [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.044651] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386970, 'name': CreateVM_Task, 'duration_secs': 0.69748} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.044828] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07899178-b53c-4b0e-877d-11774286212f] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 820.045285] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.045452] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.045782] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 820.046088] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ae9e6a2-88ce-4a2d-ba3b-c9283301bb3e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.050371] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 820.050371] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5242c9fa-716c-a11b-97b8-f12fb2c9b6c5" [ 820.050371] env[62183]: _type = "Task" [ 820.050371] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.058236] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5242c9fa-716c-a11b-97b8-f12fb2c9b6c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.117978] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d746dbb-3749-4a0f-96cf-bc78733623c2 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "27ba0765-152b-4985-927d-99670818a0b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.071s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.446019] env[62183]: DEBUG nova.scheduler.client.report [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 820.523917] env[62183]: DEBUG nova.compute.manager [req-2357d704-dc8f-4f00-855a-3eb56392d169 req-31a24ed5-671d-4731-9655-dcc64afca370 service nova] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Received event network-changed-abf57cfb-a5e5-4710-8eeb-bec026db61f5 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.524127] env[62183]: DEBUG nova.compute.manager [req-2357d704-dc8f-4f00-855a-3eb56392d169 req-31a24ed5-671d-4731-9655-dcc64afca370 service nova] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Refreshing instance network info cache due to event network-changed-abf57cfb-a5e5-4710-8eeb-bec026db61f5. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 820.524387] env[62183]: DEBUG oslo_concurrency.lockutils [req-2357d704-dc8f-4f00-855a-3eb56392d169 req-31a24ed5-671d-4731-9655-dcc64afca370 service nova] Acquiring lock "refresh_cache-0dd90b88-ed92-4331-98c5-c7481ac1ae15" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.524534] env[62183]: DEBUG oslo_concurrency.lockutils [req-2357d704-dc8f-4f00-855a-3eb56392d169 req-31a24ed5-671d-4731-9655-dcc64afca370 service nova] Acquired lock "refresh_cache-0dd90b88-ed92-4331-98c5-c7481ac1ae15" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.524914] env[62183]: DEBUG nova.network.neutron [req-2357d704-dc8f-4f00-855a-3eb56392d169 req-31a24ed5-671d-4731-9655-dcc64afca370 service nova] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Refreshing network info cache for port abf57cfb-a5e5-4710-8eeb-bec026db61f5 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.561904] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5242c9fa-716c-a11b-97b8-f12fb2c9b6c5, 'name': SearchDatastore_Task, 'duration_secs': 0.00953} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.562229] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.562491] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.562726] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.562870] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.563053] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.563377] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b4e9f0e-5af8-48c2-8b35-3bb463ca3868 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.570676] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.570843] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 820.571538] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6b32968-7a65-4321-b716-175dca1b03b0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.576397] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 820.576397] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52708762-a6e4-3043-7b59-4edee5c80f8b" [ 820.576397] env[62183]: _type = "Task" [ 820.576397] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.583822] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52708762-a6e4-3043-7b59-4edee5c80f8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.620400] env[62183]: DEBUG nova.compute.manager [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 820.704147] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.737277] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.737552] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.737708] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.737887] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.738052] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.738209] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.738415] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.738569] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.738730] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.738889] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.739067] env[62183]: DEBUG nova.virt.hardware [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.739914] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad3f055-bc25-4871-9bbd-7208ea0653a3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.748266] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c63aca8-5951-4bfd-8614-d5a18de9ce52 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.818797] env[62183]: ERROR nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port abf57cfb-a5e5-4710-8eeb-bec026db61f5, please check neutron logs for more information. [ 820.818797] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 820.818797] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.818797] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 820.818797] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 820.818797] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 820.818797] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 820.818797] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 820.818797] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.818797] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 820.818797] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.818797] env[62183]: ERROR nova.compute.manager raise self.value [ 820.818797] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 820.818797] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 820.818797] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.818797] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 820.819204] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.819204] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 820.819204] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port abf57cfb-a5e5-4710-8eeb-bec026db61f5, please check neutron logs for more information. [ 820.819204] env[62183]: ERROR nova.compute.manager [ 820.819204] env[62183]: Traceback (most recent call last): [ 820.819204] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 820.819204] env[62183]: listener.cb(fileno) [ 820.819204] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.819204] env[62183]: result = function(*args, **kwargs) [ 820.819204] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 820.819204] env[62183]: return func(*args, **kwargs) [ 820.819204] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 820.819204] env[62183]: raise e [ 820.819204] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.819204] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 820.819204] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 820.819204] env[62183]: created_port_ids = self._update_ports_for_instance( [ 820.819204] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 820.819204] env[62183]: with excutils.save_and_reraise_exception(): [ 820.819204] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.819204] env[62183]: self.force_reraise() [ 820.819204] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.819204] env[62183]: raise self.value [ 820.819204] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 820.819204] env[62183]: updated_port = self._update_port( [ 820.819204] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.819204] env[62183]: _ensure_no_port_binding_failure(port) [ 820.819204] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.819204] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 820.819832] env[62183]: nova.exception.PortBindingFailed: Binding failed for port abf57cfb-a5e5-4710-8eeb-bec026db61f5, please check neutron logs for more information. [ 820.819832] env[62183]: Removing descriptor: 16 [ 820.819832] env[62183]: ERROR nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port abf57cfb-a5e5-4710-8eeb-bec026db61f5, please check neutron logs for more information. [ 820.819832] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Traceback (most recent call last): [ 820.819832] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 820.819832] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] yield resources [ 820.819832] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 820.819832] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self.driver.spawn(context, instance, image_meta, [ 820.819832] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 820.819832] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 820.819832] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 820.819832] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] vm_ref = self.build_virtual_machine(instance, [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] vif_infos = vmwarevif.get_vif_info(self._session, [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] for vif in network_info: [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] return self._sync_wrapper(fn, *args, **kwargs) [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self.wait() [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self[:] = self._gt.wait() [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] return self._exit_event.wait() [ 820.820139] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] result = hub.switch() [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] return self.greenlet.switch() [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] result = function(*args, **kwargs) [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] return func(*args, **kwargs) [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] raise e [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] nwinfo = self.network_api.allocate_for_instance( [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 820.820442] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] created_port_ids = self._update_ports_for_instance( [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] with excutils.save_and_reraise_exception(): [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self.force_reraise() [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] raise self.value [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] updated_port = self._update_port( [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] _ensure_no_port_binding_failure(port) [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.820726] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] raise exception.PortBindingFailed(port_id=port['id']) [ 820.820978] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] nova.exception.PortBindingFailed: Binding failed for port abf57cfb-a5e5-4710-8eeb-bec026db61f5, please check neutron logs for more information. [ 820.820978] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] [ 820.820978] env[62183]: INFO nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Terminating instance [ 820.822154] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Acquiring lock "refresh_cache-0dd90b88-ed92-4331-98c5-c7481ac1ae15" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.954285] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.955111] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 820.958155] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.082s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.041603] env[62183]: DEBUG nova.network.neutron [req-2357d704-dc8f-4f00-855a-3eb56392d169 req-31a24ed5-671d-4731-9655-dcc64afca370 service nova] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.092236] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52708762-a6e4-3043-7b59-4edee5c80f8b, 'name': SearchDatastore_Task, 'duration_secs': 0.00785} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.093047] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d08be0c-c104-45ef-a4c3-4d56e9af35fb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.098481] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 821.098481] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]523c691a-281d-44e7-f31f-d40bbfbe2675" [ 821.098481] env[62183]: _type = "Task" [ 821.098481] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.106161] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]523c691a-281d-44e7-f31f-d40bbfbe2675, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.119920] env[62183]: DEBUG nova.network.neutron [req-2357d704-dc8f-4f00-855a-3eb56392d169 req-31a24ed5-671d-4731-9655-dcc64afca370 service nova] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.142714] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.462335] env[62183]: DEBUG nova.compute.utils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 821.466912] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 821.467143] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 821.520923] env[62183]: DEBUG nova.policy [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59edae5aed2e4ccbb980bc0973890baf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '594f9c41496146c68921dad4c8d99ec4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 821.613098] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]523c691a-281d-44e7-f31f-d40bbfbe2675, 'name': SearchDatastore_Task, 'duration_secs': 0.009909} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.613389] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.613650] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 07899178-b53c-4b0e-877d-11774286212f/07899178-b53c-4b0e-877d-11774286212f.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 821.613905] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f60585f-e6f6-411d-8889-8e7b2142ede5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.622776] env[62183]: DEBUG oslo_concurrency.lockutils [req-2357d704-dc8f-4f00-855a-3eb56392d169 req-31a24ed5-671d-4731-9655-dcc64afca370 service nova] Releasing lock "refresh_cache-0dd90b88-ed92-4331-98c5-c7481ac1ae15" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.623745] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 821.623745] env[62183]: value = "task-1386971" [ 821.623745] env[62183]: _type = "Task" [ 821.623745] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.623745] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Acquired lock "refresh_cache-0dd90b88-ed92-4331-98c5-c7481ac1ae15" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.623745] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.635114] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386971, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.652024] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408b7ae5-7e49-4712-9f4f-a3fbe154b757 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.660934] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba933a10-0094-48b7-b2f3-db4131d57973 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.691526] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7d0a86-09cf-4a34-a98b-27ea359d203f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.700168] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56926bc7-f8d3-413f-a78a-b3659b11915b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.712310] env[62183]: DEBUG nova.compute.provider_tree [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.959846] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Successfully created port: 8af4e5da-0239-4399-bc9e-79525f73ce64 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.970845] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.139995] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386971, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.152894] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.215593] env[62183]: DEBUG nova.scheduler.client.report [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 822.313790] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.555995] env[62183]: DEBUG nova.compute.manager [req-48ac802a-1a81-4367-8f94-ad6cda9f5193 req-ce0deb8f-3223-4d9d-99a7-ee2201bf90e3 service nova] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Received event network-vif-deleted-abf57cfb-a5e5-4710-8eeb-bec026db61f5 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.638696] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386971, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561399} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.638958] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 07899178-b53c-4b0e-877d-11774286212f/07899178-b53c-4b0e-877d-11774286212f.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 822.639268] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.639533] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df5d0281-e91c-4d15-89e3-9ec48f3962de {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.646616] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 822.646616] env[62183]: value = "task-1386972" [ 822.646616] env[62183]: _type = "Task" [ 822.646616] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.663495] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386972, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.708537] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.708840] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.720204] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.762s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.720847] env[62183]: ERROR nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f952cef-2a08-44ae-8f53-841a7833255f, please check neutron logs for more information. [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Traceback (most recent call last): [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self.driver.spawn(context, instance, image_meta, [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] vm_ref = self.build_virtual_machine(instance, [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] vif_infos = vmwarevif.get_vif_info(self._session, [ 822.720847] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] for vif in network_info: [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] return self._sync_wrapper(fn, *args, **kwargs) [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self.wait() [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self[:] = self._gt.wait() [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] return self._exit_event.wait() [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] current.throw(*self._exc) [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 822.721353] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] result = function(*args, **kwargs) [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] return func(*args, **kwargs) [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] raise e [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] nwinfo = self.network_api.allocate_for_instance( [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] created_port_ids = self._update_ports_for_instance( [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] with excutils.save_and_reraise_exception(): [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] self.force_reraise() [ 822.721820] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 822.722266] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] raise self.value [ 822.722266] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 822.722266] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] updated_port = self._update_port( [ 822.722266] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 822.722266] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] _ensure_no_port_binding_failure(port) [ 822.722266] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 822.722266] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] raise exception.PortBindingFailed(port_id=port['id']) [ 822.722266] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] nova.exception.PortBindingFailed: Binding failed for port 8f952cef-2a08-44ae-8f53-841a7833255f, please check neutron logs for more information. [ 822.722266] env[62183]: ERROR nova.compute.manager [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] [ 822.722266] env[62183]: DEBUG nova.compute.utils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Binding failed for port 8f952cef-2a08-44ae-8f53-841a7833255f, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 822.722746] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.097s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.724241] env[62183]: INFO nova.compute.claims [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.727294] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Build of instance 5a559b13-b1e0-414a-896a-e70648ed93d9 was re-scheduled: Binding failed for port 8f952cef-2a08-44ae-8f53-841a7833255f, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 822.727986] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 822.728241] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Acquiring lock "refresh_cache-5a559b13-b1e0-414a-896a-e70648ed93d9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.728340] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Acquired lock "refresh_cache-5a559b13-b1e0-414a-896a-e70648ed93d9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.728516] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 822.816023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Releasing lock "refresh_cache-0dd90b88-ed92-4331-98c5-c7481ac1ae15" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.816485] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 822.816714] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.816998] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-61d2d2c2-6f3b-4844-8e71-def4a878f9ff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.826467] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4009e36d-89df-42dc-89fd-a91d94c09fd9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.848409] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0dd90b88-ed92-4331-98c5-c7481ac1ae15 could not be found. [ 822.848576] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 822.848749] env[62183]: INFO nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Took 0.03 seconds to destroy the instance on the hypervisor. [ 822.848983] env[62183]: DEBUG oslo.service.loopingcall [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.849217] env[62183]: DEBUG nova.compute.manager [-] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.849313] env[62183]: DEBUG nova.network.neutron [-] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 822.865128] env[62183]: DEBUG nova.network.neutron [-] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.980815] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.008323] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.008568] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.008742] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.009559] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.009559] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.009559] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.009559] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.009670] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.009827] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.009926] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.010106] env[62183]: DEBUG nova.virt.hardware [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.010947] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0ab2c7-03c5-4148-9675-d454ea46725c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.014372] env[62183]: ERROR nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8af4e5da-0239-4399-bc9e-79525f73ce64, please check neutron logs for more information. [ 823.014372] env[62183]: ERROR nova.compute.manager Traceback (most recent call last): [ 823.014372] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.014372] env[62183]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 823.014372] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 823.014372] env[62183]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 823.014372] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 823.014372] env[62183]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 823.014372] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.014372] env[62183]: ERROR nova.compute.manager self.force_reraise() [ 823.014372] env[62183]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.014372] env[62183]: ERROR nova.compute.manager raise self.value [ 823.014372] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 823.014372] env[62183]: ERROR nova.compute.manager updated_port = self._update_port( [ 823.014372] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.014372] env[62183]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 823.015061] env[62183]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.015061] env[62183]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 823.015061] env[62183]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8af4e5da-0239-4399-bc9e-79525f73ce64, please check neutron logs for more information. [ 823.015061] env[62183]: ERROR nova.compute.manager [ 823.015061] env[62183]: Traceback (most recent call last): [ 823.015061] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 823.015061] env[62183]: listener.cb(fileno) [ 823.015061] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 823.015061] env[62183]: result = function(*args, **kwargs) [ 823.015061] env[62183]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 823.015061] env[62183]: return func(*args, **kwargs) [ 823.015061] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 823.015061] env[62183]: raise e [ 823.015061] env[62183]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.015061] env[62183]: nwinfo = self.network_api.allocate_for_instance( [ 823.015061] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 823.015061] env[62183]: created_port_ids = self._update_ports_for_instance( [ 823.015061] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 823.015061] env[62183]: with excutils.save_and_reraise_exception(): [ 823.015061] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.015061] env[62183]: self.force_reraise() [ 823.015061] env[62183]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.015061] env[62183]: raise self.value [ 823.015061] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 823.015061] env[62183]: updated_port = self._update_port( [ 823.015061] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.015061] env[62183]: _ensure_no_port_binding_failure(port) [ 823.015061] env[62183]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.015061] env[62183]: raise exception.PortBindingFailed(port_id=port['id']) [ 823.016162] env[62183]: nova.exception.PortBindingFailed: Binding failed for port 8af4e5da-0239-4399-bc9e-79525f73ce64, please check neutron logs for more information. [ 823.016162] env[62183]: Removing descriptor: 16 [ 823.020198] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4ce088-4925-46f3-a04f-73de313698bf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.035621] env[62183]: ERROR nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8af4e5da-0239-4399-bc9e-79525f73ce64, please check neutron logs for more information. [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Traceback (most recent call last): [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] yield resources [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self.driver.spawn(context, instance, image_meta, [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] vm_ref = self.build_virtual_machine(instance, [ 823.035621] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] for vif in network_info: [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] return self._sync_wrapper(fn, *args, **kwargs) [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self.wait() [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self[:] = self._gt.wait() [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] return self._exit_event.wait() [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 823.036032] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] current.throw(*self._exc) [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] result = function(*args, **kwargs) [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] return func(*args, **kwargs) [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] raise e [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] nwinfo = self.network_api.allocate_for_instance( [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] created_port_ids = self._update_ports_for_instance( [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] with excutils.save_and_reraise_exception(): [ 823.036363] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self.force_reraise() [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] raise self.value [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] updated_port = self._update_port( [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] _ensure_no_port_binding_failure(port) [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] raise exception.PortBindingFailed(port_id=port['id']) [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] nova.exception.PortBindingFailed: Binding failed for port 8af4e5da-0239-4399-bc9e-79525f73ce64, please check neutron logs for more information. [ 823.036712] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] [ 823.036712] env[62183]: INFO nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Terminating instance [ 823.038025] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.038183] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.038349] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.157523] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386972, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067703} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.157806] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.158574] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3a4e3b-4ceb-4910-b7f0-66022acd0c00 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.178139] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 07899178-b53c-4b0e-877d-11774286212f/07899178-b53c-4b0e-877d-11774286212f.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.178742] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01077100-40d6-4602-b344-1f38fbe2084a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.197283] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 823.197283] env[62183]: value = "task-1386973" [ 823.197283] env[62183]: _type = "Task" [ 823.197283] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.204724] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386973, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.249285] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.340998] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.367616] env[62183]: DEBUG nova.network.neutron [-] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.555810] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.630328] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.706435] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386973, 'name': ReconfigVM_Task, 'duration_secs': 0.309194} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.706743] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 07899178-b53c-4b0e-877d-11774286212f/07899178-b53c-4b0e-877d-11774286212f.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.707345] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-894ad925-59bd-4b52-800c-e2f1dc5d1cbe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.713630] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 823.713630] env[62183]: value = "task-1386974" [ 823.713630] env[62183]: _type = "Task" [ 823.713630] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.721128] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386974, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.843821] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Releasing lock "refresh_cache-5a559b13-b1e0-414a-896a-e70648ed93d9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.844077] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 823.844318] env[62183]: DEBUG nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.844503] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.860852] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.871713] env[62183]: INFO nova.compute.manager [-] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Took 1.02 seconds to deallocate network for instance. [ 823.874126] env[62183]: DEBUG nova.compute.claims [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 823.874377] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.900189] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b44c1c-b2ba-4e41-a768-b9df8097d992 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.907784] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4573ef74-5df5-4bb2-b3f1-b7e1e6674f5a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.942540] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c0f25d-9826-41dc-8d72-88622dc8042b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.950119] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55c7471-1972-4001-aac8-a128c5706812 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.963682] env[62183]: DEBUG nova.compute.provider_tree [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.133621] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.134107] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 824.134399] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 824.134720] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-272bd060-fcb6-49b2-b968-149049873eea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.143728] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715753cc-f782-4800-9d3b-be1d49d3719b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.171879] env[62183]: WARNING nova.virt.vmwareapi.vmops [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4 could not be found. [ 824.172090] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 824.172282] env[62183]: INFO nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 824.172529] env[62183]: DEBUG oslo.service.loopingcall [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.172752] env[62183]: DEBUG nova.compute.manager [-] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.172845] env[62183]: DEBUG nova.network.neutron [-] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 824.190956] env[62183]: DEBUG nova.network.neutron [-] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.227730] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386974, 'name': Rename_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.365440] env[62183]: DEBUG nova.network.neutron [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.470135] env[62183]: DEBUG nova.scheduler.client.report [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 824.586237] env[62183]: DEBUG nova.compute.manager [req-bdde33dc-62d4-4869-bc9f-a821da75edcd req-62ae992c-c0ea-4659-9a74-732cf7ffd5b0 service nova] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Received event network-changed-8af4e5da-0239-4399-bc9e-79525f73ce64 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.586237] env[62183]: DEBUG nova.compute.manager [req-bdde33dc-62d4-4869-bc9f-a821da75edcd req-62ae992c-c0ea-4659-9a74-732cf7ffd5b0 service nova] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Refreshing instance network info cache due to event network-changed-8af4e5da-0239-4399-bc9e-79525f73ce64. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 824.586237] env[62183]: DEBUG oslo_concurrency.lockutils [req-bdde33dc-62d4-4869-bc9f-a821da75edcd req-62ae992c-c0ea-4659-9a74-732cf7ffd5b0 service nova] Acquiring lock "refresh_cache-6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.586237] env[62183]: DEBUG oslo_concurrency.lockutils [req-bdde33dc-62d4-4869-bc9f-a821da75edcd req-62ae992c-c0ea-4659-9a74-732cf7ffd5b0 service nova] Acquired lock "refresh_cache-6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.586667] env[62183]: DEBUG nova.network.neutron [req-bdde33dc-62d4-4869-bc9f-a821da75edcd req-62ae992c-c0ea-4659-9a74-732cf7ffd5b0 service nova] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Refreshing network info cache for port 8af4e5da-0239-4399-bc9e-79525f73ce64 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.697338] env[62183]: DEBUG nova.network.neutron [-] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.724539] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386974, 'name': Rename_Task, 'duration_secs': 0.93562} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.726023] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 824.726023] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b763909-6656-492c-9c0e-a1b93c1370be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.731543] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 824.731543] env[62183]: value = "task-1386975" [ 824.731543] env[62183]: _type = "Task" [ 824.731543] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.741789] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386975, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.870621] env[62183]: INFO nova.compute.manager [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] [instance: 5a559b13-b1e0-414a-896a-e70648ed93d9] Took 1.02 seconds to deallocate network for instance. [ 824.973617] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.974677] env[62183]: DEBUG nova.compute.manager [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 824.978288] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.665s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.979723] env[62183]: INFO nova.compute.claims [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.105569] env[62183]: DEBUG nova.network.neutron [req-bdde33dc-62d4-4869-bc9f-a821da75edcd req-62ae992c-c0ea-4659-9a74-732cf7ffd5b0 service nova] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.191785] env[62183]: DEBUG nova.network.neutron [req-bdde33dc-62d4-4869-bc9f-a821da75edcd req-62ae992c-c0ea-4659-9a74-732cf7ffd5b0 service nova] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.198682] env[62183]: INFO nova.compute.manager [-] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Took 1.03 seconds to deallocate network for instance. [ 825.201046] env[62183]: DEBUG nova.compute.claims [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Aborting claim: {{(pid=62183) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 825.201226] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.241982] env[62183]: DEBUG oslo_vmware.api [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386975, 'name': PowerOnVM_Task, 'duration_secs': 0.419658} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.242783] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.242783] env[62183]: DEBUG nova.compute.manager [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.243423] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ec839c-9e2b-4dd9-9ecd-2aad56347968 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.480384] env[62183]: DEBUG nova.compute.utils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.481757] env[62183]: DEBUG nova.compute.manager [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 825.481930] env[62183]: DEBUG nova.network.neutron [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 825.532412] env[62183]: DEBUG nova.policy [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be747fe0622548afa03792bb0a46c726', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c9dfd905ec14f2bbbe2b98692222f17', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 825.694466] env[62183]: DEBUG oslo_concurrency.lockutils [req-bdde33dc-62d4-4869-bc9f-a821da75edcd req-62ae992c-c0ea-4659-9a74-732cf7ffd5b0 service nova] Releasing lock "refresh_cache-6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.694830] env[62183]: DEBUG nova.compute.manager [req-bdde33dc-62d4-4869-bc9f-a821da75edcd req-62ae992c-c0ea-4659-9a74-732cf7ffd5b0 service nova] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Received event network-vif-deleted-8af4e5da-0239-4399-bc9e-79525f73ce64 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.759926] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.815064] env[62183]: DEBUG nova.network.neutron [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Successfully created port: dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.895804] env[62183]: INFO nova.scheduler.client.report [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Deleted allocations for instance 5a559b13-b1e0-414a-896a-e70648ed93d9 [ 825.984632] env[62183]: DEBUG nova.compute.manager [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.100590] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "07899178-b53c-4b0e-877d-11774286212f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.100590] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "07899178-b53c-4b0e-877d-11774286212f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.100590] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "07899178-b53c-4b0e-877d-11774286212f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.100744] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "07899178-b53c-4b0e-877d-11774286212f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.100999] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "07899178-b53c-4b0e-877d-11774286212f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.103770] env[62183]: INFO nova.compute.manager [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Terminating instance [ 826.105381] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "refresh_cache-07899178-b53c-4b0e-877d-11774286212f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.105537] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquired lock "refresh_cache-07899178-b53c-4b0e-877d-11774286212f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.105715] env[62183]: DEBUG nova.network.neutron [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.161756] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac7d530-a99c-4682-9929-369c00c05a03 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.169811] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8d283d-8d40-48c0-b541-e50f24964391 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.201120] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97230995-7f8c-47b0-9f6c-03077cc7f39c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.208398] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1f9d05-a7bb-4e27-a52e-078cb885856f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.223576] env[62183]: DEBUG nova.compute.provider_tree [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.403125] env[62183]: DEBUG oslo_concurrency.lockutils [None req-69752ac9-5ccf-482f-87e7-0fc8aafafef0 tempest-ServerActionsV293TestJSON-2012225577 tempest-ServerActionsV293TestJSON-2012225577-project-member] Lock "5a559b13-b1e0-414a-896a-e70648ed93d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.853s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.626294] env[62183]: DEBUG nova.network.neutron [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.691625] env[62183]: DEBUG nova.network.neutron [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.727265] env[62183]: DEBUG nova.scheduler.client.report [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 826.905625] env[62183]: DEBUG nova.compute.manager [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.996375] env[62183]: DEBUG nova.compute.manager [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.025879] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.026168] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.026325] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.026506] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.026692] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.026858] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.027084] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.027246] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.027414] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.027572] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.027753] env[62183]: DEBUG nova.virt.hardware [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.028635] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f295542-4a3b-47e6-8753-d2f691c1263d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.037110] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0788eb-f97e-4b34-bd84-7783f915714e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.197023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Releasing lock "refresh_cache-07899178-b53c-4b0e-877d-11774286212f" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.197023] env[62183]: DEBUG nova.compute.manager [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.197023] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 827.197023] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a01413-1ccc-4d8e-a8da-d1ccf6fad0d7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.208160] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 827.208699] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8492065a-1de0-4c28-ba9f-c626b1fa298c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.218024] env[62183]: DEBUG oslo_vmware.api [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 827.218024] env[62183]: value = "task-1386977" [ 827.218024] env[62183]: _type = "Task" [ 827.218024] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.230715] env[62183]: DEBUG oslo_vmware.api [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386977, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.231952] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.254s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.232874] env[62183]: DEBUG nova.compute.manager [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.239106] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.058s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.239106] env[62183]: INFO nova.compute.claims [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.384893] env[62183]: DEBUG nova.compute.manager [req-fc55a7f7-69c1-4d62-9b06-c6e84a4c7faf req-067fc9c2-25b5-4e52-931c-c857efe163ea service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Received event network-vif-plugged-dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.385153] env[62183]: DEBUG oslo_concurrency.lockutils [req-fc55a7f7-69c1-4d62-9b06-c6e84a4c7faf req-067fc9c2-25b5-4e52-931c-c857efe163ea service nova] Acquiring lock "7465df88-c68c-49d3-9a91-6fff0d06957b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.385406] env[62183]: DEBUG oslo_concurrency.lockutils [req-fc55a7f7-69c1-4d62-9b06-c6e84a4c7faf req-067fc9c2-25b5-4e52-931c-c857efe163ea service nova] Lock "7465df88-c68c-49d3-9a91-6fff0d06957b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.385577] env[62183]: DEBUG oslo_concurrency.lockutils [req-fc55a7f7-69c1-4d62-9b06-c6e84a4c7faf req-067fc9c2-25b5-4e52-931c-c857efe163ea service nova] Lock "7465df88-c68c-49d3-9a91-6fff0d06957b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.385769] env[62183]: DEBUG nova.compute.manager [req-fc55a7f7-69c1-4d62-9b06-c6e84a4c7faf req-067fc9c2-25b5-4e52-931c-c857efe163ea service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] No waiting events found dispatching network-vif-plugged-dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 827.385934] env[62183]: WARNING nova.compute.manager [req-fc55a7f7-69c1-4d62-9b06-c6e84a4c7faf req-067fc9c2-25b5-4e52-931c-c857efe163ea service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Received unexpected event network-vif-plugged-dccdd9b2-d8fa-435b-86d8-ec505124174c for instance with vm_state building and task_state spawning. [ 827.429482] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.727466] env[62183]: DEBUG oslo_vmware.api [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386977, 'name': PowerOffVM_Task, 'duration_secs': 0.129495} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.727766] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 827.727926] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 827.728365] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b19faf5-4e06-425b-a12e-8337b9f0e59a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.744184] env[62183]: DEBUG nova.compute.utils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.747551] env[62183]: DEBUG nova.compute.manager [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.747809] env[62183]: DEBUG nova.network.neutron [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 827.751792] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 827.751981] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 827.752178] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Deleting the datastore file [datastore1] 07899178-b53c-4b0e-877d-11774286212f {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.752431] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e4c335e-19e6-4115-8426-0e3d3507e4af {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.758143] env[62183]: DEBUG oslo_vmware.api [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for the task: (returnval){ [ 827.758143] env[62183]: value = "task-1386979" [ 827.758143] env[62183]: _type = "Task" [ 827.758143] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.766129] env[62183]: DEBUG oslo_vmware.api [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386979, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.815989] env[62183]: DEBUG nova.policy [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e0eee67c7954c80b4e1ea43b0abbcb5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebcc716f958942b588a6bfde78d2c00d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 828.075198] env[62183]: DEBUG nova.network.neutron [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Successfully updated port: dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.112261] env[62183]: DEBUG nova.compute.manager [req-7652a2b0-b5a2-4f3b-ae82-cd4bbb8cc103 req-5743ca86-4440-489d-91ad-61afaed8748f service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Received event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.112444] env[62183]: DEBUG nova.compute.manager [req-7652a2b0-b5a2-4f3b-ae82-cd4bbb8cc103 req-5743ca86-4440-489d-91ad-61afaed8748f service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing instance network info cache due to event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 828.112781] env[62183]: DEBUG oslo_concurrency.lockutils [req-7652a2b0-b5a2-4f3b-ae82-cd4bbb8cc103 req-5743ca86-4440-489d-91ad-61afaed8748f service nova] Acquiring lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.112781] env[62183]: DEBUG oslo_concurrency.lockutils [req-7652a2b0-b5a2-4f3b-ae82-cd4bbb8cc103 req-5743ca86-4440-489d-91ad-61afaed8748f service nova] Acquired lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.113178] env[62183]: DEBUG nova.network.neutron [req-7652a2b0-b5a2-4f3b-ae82-cd4bbb8cc103 req-5743ca86-4440-489d-91ad-61afaed8748f service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing network info cache for port dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.248462] env[62183]: DEBUG nova.compute.manager [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.269088] env[62183]: DEBUG oslo_vmware.api [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Task: {'id': task-1386979, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103051} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.269383] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 828.269571] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 828.269742] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 828.269915] env[62183]: INFO nova.compute.manager [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Took 1.07 seconds to destroy the instance on the hypervisor. [ 828.270182] env[62183]: DEBUG oslo.service.loopingcall [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.270378] env[62183]: DEBUG nova.compute.manager [-] [instance: 07899178-b53c-4b0e-877d-11774286212f] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 828.270472] env[62183]: DEBUG nova.network.neutron [-] [instance: 07899178-b53c-4b0e-877d-11774286212f] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 828.297837] env[62183]: DEBUG nova.network.neutron [-] [instance: 07899178-b53c-4b0e-877d-11774286212f] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.485660] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1aa465-1ae6-4c97-a771-6db7241a1713 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.495248] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edf2665-226a-488f-ace6-28db3cd6d977 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.529812] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20bace92-b348-4fa2-b4c6-0b55242546d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.539218] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8fd375-2d6e-4442-917b-dacf79c9c3c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.542398] env[62183]: DEBUG nova.network.neutron [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Successfully created port: 91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.551889] env[62183]: DEBUG nova.compute.provider_tree [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.576581] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.684543] env[62183]: DEBUG nova.network.neutron [req-7652a2b0-b5a2-4f3b-ae82-cd4bbb8cc103 req-5743ca86-4440-489d-91ad-61afaed8748f service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.800726] env[62183]: DEBUG nova.network.neutron [-] [instance: 07899178-b53c-4b0e-877d-11774286212f] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.810114] env[62183]: DEBUG nova.network.neutron [req-7652a2b0-b5a2-4f3b-ae82-cd4bbb8cc103 req-5743ca86-4440-489d-91ad-61afaed8748f service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.055383] env[62183]: DEBUG nova.scheduler.client.report [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 829.262450] env[62183]: DEBUG nova.compute.manager [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.287863] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.288119] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.288275] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.288455] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.288597] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.288738] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.288937] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.289103] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.289267] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.289424] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.289593] env[62183]: DEBUG nova.virt.hardware [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.290469] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0bbecc4-a591-4704-afee-2583171456e8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.297882] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60780937-2a5b-432d-a702-69c96f016f59 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.303384] env[62183]: INFO nova.compute.manager [-] [instance: 07899178-b53c-4b0e-877d-11774286212f] Took 1.03 seconds to deallocate network for instance. [ 829.317634] env[62183]: DEBUG oslo_concurrency.lockutils [req-7652a2b0-b5a2-4f3b-ae82-cd4bbb8cc103 req-5743ca86-4440-489d-91ad-61afaed8748f service nova] Releasing lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.320152] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquired lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.320308] env[62183]: DEBUG nova.network.neutron [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 829.559962] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.560520] env[62183]: DEBUG nova.compute.manager [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.563041] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.339s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.564492] env[62183]: INFO nova.compute.claims [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.822722] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.847914] env[62183]: DEBUG nova.network.neutron [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.068661] env[62183]: DEBUG nova.compute.utils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.076023] env[62183]: DEBUG nova.compute.manager [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.076023] env[62183]: DEBUG nova.network.neutron [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 830.080264] env[62183]: DEBUG nova.network.neutron [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updating instance_info_cache with network_info: [{"id": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "address": "fa:16:3e:f8:f1:0b", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdccdd9b2-d8", "ovs_interfaceid": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.114769] env[62183]: DEBUG nova.policy [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c4bbb2b919d416aaff771b5c9b2cc04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '072024d50c3b49d09a721a3e3b7fd856', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.268790] env[62183]: DEBUG nova.compute.manager [req-27ffdaac-e748-4e9a-9e44-d5676f684332 req-d3d42606-0776-4b35-985d-370573854f5f service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received event network-vif-plugged-91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.269061] env[62183]: DEBUG oslo_concurrency.lockutils [req-27ffdaac-e748-4e9a-9e44-d5676f684332 req-d3d42606-0776-4b35-985d-370573854f5f service nova] Acquiring lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.269215] env[62183]: DEBUG oslo_concurrency.lockutils [req-27ffdaac-e748-4e9a-9e44-d5676f684332 req-d3d42606-0776-4b35-985d-370573854f5f service nova] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.269383] env[62183]: DEBUG oslo_concurrency.lockutils [req-27ffdaac-e748-4e9a-9e44-d5676f684332 req-d3d42606-0776-4b35-985d-370573854f5f service nova] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.269588] env[62183]: DEBUG nova.compute.manager [req-27ffdaac-e748-4e9a-9e44-d5676f684332 req-d3d42606-0776-4b35-985d-370573854f5f service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] No waiting events found dispatching network-vif-plugged-91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 830.269747] env[62183]: WARNING nova.compute.manager [req-27ffdaac-e748-4e9a-9e44-d5676f684332 req-d3d42606-0776-4b35-985d-370573854f5f service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received unexpected event network-vif-plugged-91e1877d-c4f8-4ad7-8076-08c985c66e4c for instance with vm_state building and task_state spawning. [ 830.301365] env[62183]: DEBUG nova.network.neutron [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Successfully updated port: 91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 830.400246] env[62183]: DEBUG nova.network.neutron [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Successfully created port: 1da5742b-a895-4c96-88f1-c84646ab800c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.576758] env[62183]: DEBUG nova.compute.manager [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.582429] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Releasing lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.582429] env[62183]: DEBUG nova.compute.manager [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Instance network_info: |[{"id": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "address": "fa:16:3e:f8:f1:0b", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdccdd9b2-d8", "ovs_interfaceid": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.583909] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:f1:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0721b358-3768-472d-95f8-6d6755ab1635', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dccdd9b2-d8fa-435b-86d8-ec505124174c', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.591239] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Creating folder: Project (4c9dfd905ec14f2bbbe2b98692222f17). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.591586] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f45eae4b-a946-4146-9e39-4a10616e11f1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.604332] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Created folder: Project (4c9dfd905ec14f2bbbe2b98692222f17) in parent group-v294392. [ 830.604522] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Creating folder: Instances. Parent ref: group-v294421. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.604943] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc5c3f83-1ee9-458f-901d-8ad90b1995bf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.615330] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Created folder: Instances in parent group-v294421. [ 830.615703] env[62183]: DEBUG oslo.service.loopingcall [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.619532] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.620768] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8eaa9751-c4c5-4f27-a820-34fe5a2cb236 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.644662] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.644662] env[62183]: value = "task-1386982" [ 830.644662] env[62183]: _type = "Task" [ 830.644662] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.653096] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386982, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.786590] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98813765-5983-4f53-b564-bd779c03cf0a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.794205] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cda2b9f-07fc-4166-8142-e6d0fa09f339 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.824394] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.824523] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.824671] env[62183]: DEBUG nova.network.neutron [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 830.826420] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ff1b1c-d4e1-4aeb-a611-fce1d4a9f2d9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.834226] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8987134-7d79-421b-94e1-6acf710bc8b7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.847342] env[62183]: DEBUG nova.compute.provider_tree [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.154035] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386982, 'name': CreateVM_Task, 'duration_secs': 0.311367} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.155030] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.161264] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.161448] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.161760] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.161997] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1daf380-6639-4d1c-a8a0-d0da77fd3f5f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.166172] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 831.166172] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52589bbb-0b10-4e58-ead9-ec62a0a3a111" [ 831.166172] env[62183]: _type = "Task" [ 831.166172] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.173118] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52589bbb-0b10-4e58-ead9-ec62a0a3a111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.350258] env[62183]: DEBUG nova.scheduler.client.report [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 831.358182] env[62183]: DEBUG nova.network.neutron [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 831.527385] env[62183]: DEBUG nova.network.neutron [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.586743] env[62183]: DEBUG nova.compute.manager [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.613085] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.613348] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.613545] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.613734] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.613880] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.614037] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.614244] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.614402] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.614566] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.614724] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.614894] env[62183]: DEBUG nova.virt.hardware [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.615783] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df80ac96-e696-4851-bbbf-d8c6ac7837ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.623485] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e3a831-72da-4c51-b7e9-d1baead6aa6c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.677704] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52589bbb-0b10-4e58-ead9-ec62a0a3a111, 'name': SearchDatastore_Task, 'duration_secs': 0.008965} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.678388] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.678720] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.680242] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.680242] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.680242] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.680242] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6043f7fa-d910-4e51-b38a-261b3baf937b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.690253] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.690430] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 831.691253] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e62322c-5e48-4866-b6f1-5eaa0d856552 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.696163] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 831.696163] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52b2706e-fd57-18d3-91b2-34b1bbfd3d57" [ 831.696163] env[62183]: _type = "Task" [ 831.696163] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.707253] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b2706e-fd57-18d3-91b2-34b1bbfd3d57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.855236] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.855805] env[62183]: DEBUG nova.compute.manager [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.858774] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.276s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.860259] env[62183]: INFO nova.compute.claims [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.911894] env[62183]: DEBUG nova.compute.manager [req-1cf09aaf-75c6-495d-bc9d-aae617a0a461 req-e2bce306-e1bd-458a-a2f5-bd2cf5726b69 service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Received event network-vif-plugged-1da5742b-a895-4c96-88f1-c84646ab800c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.912573] env[62183]: DEBUG oslo_concurrency.lockutils [req-1cf09aaf-75c6-495d-bc9d-aae617a0a461 req-e2bce306-e1bd-458a-a2f5-bd2cf5726b69 service nova] Acquiring lock "68e791b4-61db-4b6b-a30a-ccf5d6657643-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.912801] env[62183]: DEBUG oslo_concurrency.lockutils [req-1cf09aaf-75c6-495d-bc9d-aae617a0a461 req-e2bce306-e1bd-458a-a2f5-bd2cf5726b69 service nova] Lock "68e791b4-61db-4b6b-a30a-ccf5d6657643-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.912971] env[62183]: DEBUG oslo_concurrency.lockutils [req-1cf09aaf-75c6-495d-bc9d-aae617a0a461 req-e2bce306-e1bd-458a-a2f5-bd2cf5726b69 service nova] Lock "68e791b4-61db-4b6b-a30a-ccf5d6657643-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.913522] env[62183]: DEBUG nova.compute.manager [req-1cf09aaf-75c6-495d-bc9d-aae617a0a461 req-e2bce306-e1bd-458a-a2f5-bd2cf5726b69 service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] No waiting events found dispatching network-vif-plugged-1da5742b-a895-4c96-88f1-c84646ab800c {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.913766] env[62183]: WARNING nova.compute.manager [req-1cf09aaf-75c6-495d-bc9d-aae617a0a461 req-e2bce306-e1bd-458a-a2f5-bd2cf5726b69 service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Received unexpected event network-vif-plugged-1da5742b-a895-4c96-88f1-c84646ab800c for instance with vm_state building and task_state spawning. [ 832.010411] env[62183]: DEBUG nova.network.neutron [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Successfully updated port: 1da5742b-a895-4c96-88f1-c84646ab800c {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.030225] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.030493] env[62183]: DEBUG nova.compute.manager [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Instance network_info: |[{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 832.031217] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:8a:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db00ec2e-3155-46b6-8170-082f7d86dbe7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91e1877d-c4f8-4ad7-8076-08c985c66e4c', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.038417] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Creating folder: Project (ebcc716f958942b588a6bfde78d2c00d). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 832.038853] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74476bdd-4800-4f73-8431-f699afd21366 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.049649] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Created folder: Project (ebcc716f958942b588a6bfde78d2c00d) in parent group-v294392. [ 832.049828] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Creating folder: Instances. Parent ref: group-v294424. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 832.050063] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c61aa9d1-d86e-483a-b41c-aa5923b89cad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.059032] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Created folder: Instances in parent group-v294424. [ 832.059317] env[62183]: DEBUG oslo.service.loopingcall [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.059538] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 832.059786] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32ae3021-147c-4cc7-83e7-a5d4dc43ed43 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.078964] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.078964] env[62183]: value = "task-1386985" [ 832.078964] env[62183]: _type = "Task" [ 832.078964] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.086618] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386985, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.207368] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b2706e-fd57-18d3-91b2-34b1bbfd3d57, 'name': SearchDatastore_Task, 'duration_secs': 0.008091} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.208251] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb1fed98-973a-4416-b23e-adced0f379ab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.213832] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 832.213832] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52804c9c-07dc-cec2-4c60-e5cba3855438" [ 832.213832] env[62183]: _type = "Task" [ 832.213832] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.223044] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52804c9c-07dc-cec2-4c60-e5cba3855438, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.296318] env[62183]: DEBUG nova.compute.manager [req-0b6bba20-865c-4059-baca-f4404168795a req-fbe6a4d7-2cdf-447f-8f3a-5237319be30e service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.296498] env[62183]: DEBUG nova.compute.manager [req-0b6bba20-865c-4059-baca-f4404168795a req-fbe6a4d7-2cdf-447f-8f3a-5237319be30e service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing instance network info cache due to event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.296712] env[62183]: DEBUG oslo_concurrency.lockutils [req-0b6bba20-865c-4059-baca-f4404168795a req-fbe6a4d7-2cdf-447f-8f3a-5237319be30e service nova] Acquiring lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.296857] env[62183]: DEBUG oslo_concurrency.lockutils [req-0b6bba20-865c-4059-baca-f4404168795a req-fbe6a4d7-2cdf-447f-8f3a-5237319be30e service nova] Acquired lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.297022] env[62183]: DEBUG nova.network.neutron [req-0b6bba20-865c-4059-baca-f4404168795a req-fbe6a4d7-2cdf-447f-8f3a-5237319be30e service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.365106] env[62183]: DEBUG nova.compute.utils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.368592] env[62183]: DEBUG nova.compute.manager [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.368772] env[62183]: DEBUG nova.network.neutron [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 832.437855] env[62183]: DEBUG nova.policy [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd763291af60146b29874a4fa5ba3ef3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '36152664258e4b0fbe576e8fbf1423ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.512828] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquiring lock "refresh_cache-68e791b4-61db-4b6b-a30a-ccf5d6657643" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.512986] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquired lock "refresh_cache-68e791b4-61db-4b6b-a30a-ccf5d6657643" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.513164] env[62183]: DEBUG nova.network.neutron [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.589149] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386985, 'name': CreateVM_Task, 'duration_secs': 0.327526} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.589403] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 832.590137] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.590320] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.591233] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.591743] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d72649fb-4325-4e2f-a629-c68daca00a84 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.596713] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 832.596713] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a38313-8be4-958a-64ac-0eb49200adb3" [ 832.596713] env[62183]: _type = "Task" [ 832.596713] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.605052] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a38313-8be4-958a-64ac-0eb49200adb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.734228] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52804c9c-07dc-cec2-4c60-e5cba3855438, 'name': SearchDatastore_Task, 'duration_secs': 0.010087} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.734228] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.734228] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 7465df88-c68c-49d3-9a91-6fff0d06957b/7465df88-c68c-49d3-9a91-6fff0d06957b.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 832.734228] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-abc71c4c-3755-4bd9-b6b0-a83524714ed6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.745023] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 832.745023] env[62183]: value = "task-1386986" [ 832.745023] env[62183]: _type = "Task" [ 832.745023] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.753210] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.754291] env[62183]: DEBUG nova.network.neutron [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Successfully created port: eb7f6201-df05-4a22-b815-57d19d54b5fa {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.869804] env[62183]: DEBUG nova.compute.manager [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.057949] env[62183]: DEBUG nova.network.neutron [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.103176] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f24625b-fbe0-4aaa-aa44-d42bac7b1252 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.112194] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a38313-8be4-958a-64ac-0eb49200adb3, 'name': SearchDatastore_Task, 'duration_secs': 0.009678} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.114639] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.116034] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.116034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.116034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.116034] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.116034] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ef6a7d1-4bbd-4769-a2a6-9b0035e4f56a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.120761] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e4c9db0-fb0e-4b9e-936c-20da1b641475 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.155136] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb799aa-972d-4e43-aba1-89c05cbce941 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.158085] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.158367] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 833.159163] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0ebccc7-dcec-49b9-996f-cece8fffb369 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.164544] env[62183]: DEBUG nova.network.neutron [req-0b6bba20-865c-4059-baca-f4404168795a req-fbe6a4d7-2cdf-447f-8f3a-5237319be30e service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updated VIF entry in instance network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 833.164995] env[62183]: DEBUG nova.network.neutron [req-0b6bba20-865c-4059-baca-f4404168795a req-fbe6a4d7-2cdf-447f-8f3a-5237319be30e service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.169533] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 833.169533] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52513c90-df49-99f4-c9f0-41422de8df4b" [ 833.169533] env[62183]: _type = "Task" [ 833.169533] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.171152] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cda7d2a-a3eb-4125-b15e-839b988edda3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.193663] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52513c90-df49-99f4-c9f0-41422de8df4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.194102] env[62183]: DEBUG nova.compute.provider_tree [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.253367] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386986, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.275944] env[62183]: DEBUG nova.network.neutron [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Updating instance_info_cache with network_info: [{"id": "1da5742b-a895-4c96-88f1-c84646ab800c", "address": "fa:16:3e:57:d8:e2", "network": {"id": "559cdc13-7bba-43ae-8ddf-6c03d4efebb8", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1367814136-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "072024d50c3b49d09a721a3e3b7fd856", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1da5742b-a8", "ovs_interfaceid": "1da5742b-a895-4c96-88f1-c84646ab800c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.671040] env[62183]: DEBUG oslo_concurrency.lockutils [req-0b6bba20-865c-4059-baca-f4404168795a req-fbe6a4d7-2cdf-447f-8f3a-5237319be30e service nova] Releasing lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.681548] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52513c90-df49-99f4-c9f0-41422de8df4b, 'name': SearchDatastore_Task, 'duration_secs': 0.059842} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.682350] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6105854c-9187-473e-b570-fb20dc9d9a3c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.688324] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 833.688324] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52bd48bb-4a37-2880-d97a-781738dc42a4" [ 833.688324] env[62183]: _type = "Task" [ 833.688324] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.696563] env[62183]: DEBUG nova.scheduler.client.report [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 833.699559] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bd48bb-4a37-2880-d97a-781738dc42a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.754959] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386986, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553274} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.755367] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 7465df88-c68c-49d3-9a91-6fff0d06957b/7465df88-c68c-49d3-9a91-6fff0d06957b.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 833.755691] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.756081] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-32de24b6-8e0e-4e61-aacb-82f63580ce86 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.763786] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 833.763786] env[62183]: value = "task-1386987" [ 833.763786] env[62183]: _type = "Task" [ 833.763786] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.774580] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386987, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.778846] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Releasing lock "refresh_cache-68e791b4-61db-4b6b-a30a-ccf5d6657643" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.779309] env[62183]: DEBUG nova.compute.manager [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Instance network_info: |[{"id": "1da5742b-a895-4c96-88f1-c84646ab800c", "address": "fa:16:3e:57:d8:e2", "network": {"id": "559cdc13-7bba-43ae-8ddf-6c03d4efebb8", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1367814136-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "072024d50c3b49d09a721a3e3b7fd856", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1da5742b-a8", "ovs_interfaceid": "1da5742b-a895-4c96-88f1-c84646ab800c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.779869] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:d8:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b399c74-1411-408a-b4cd-84e268ae83fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1da5742b-a895-4c96-88f1-c84646ab800c', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.793052] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Creating folder: Project (072024d50c3b49d09a721a3e3b7fd856). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.793192] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f620c42d-7852-4702-b924-7ce8ff5f69e9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.804452] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Created folder: Project (072024d50c3b49d09a721a3e3b7fd856) in parent group-v294392. [ 833.805837] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Creating folder: Instances. Parent ref: group-v294427. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.805837] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-80cfa6d0-f132-4a00-be86-87449c46ae42 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.814648] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Created folder: Instances in parent group-v294427. [ 833.815060] env[62183]: DEBUG oslo.service.loopingcall [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.815318] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.815631] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cabe3340-d4b4-43a2-a001-7b873f7a4bf0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.846350] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.846350] env[62183]: value = "task-1386990" [ 833.846350] env[62183]: _type = "Task" [ 833.846350] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.854444] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386990, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.884977] env[62183]: DEBUG nova.compute.manager [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.917815] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.918087] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.918244] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.918428] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.918682] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.918998] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.919260] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.919431] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.919617] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.919787] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.919964] env[62183]: DEBUG nova.virt.hardware [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.921015] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87fbffb-dfeb-419a-9fe5-a5a179c2fc50 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.929027] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591486d9-fa21-4a51-9ffd-10ed28363dcb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.958689] env[62183]: DEBUG nova.compute.manager [req-57717834-aeef-4980-8ad8-00fd7c9e5655 req-57318590-b1d3-4009-a6c4-c107b3227e01 service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Received event network-changed-1da5742b-a895-4c96-88f1-c84646ab800c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.958872] env[62183]: DEBUG nova.compute.manager [req-57717834-aeef-4980-8ad8-00fd7c9e5655 req-57318590-b1d3-4009-a6c4-c107b3227e01 service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Refreshing instance network info cache due to event network-changed-1da5742b-a895-4c96-88f1-c84646ab800c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.959053] env[62183]: DEBUG oslo_concurrency.lockutils [req-57717834-aeef-4980-8ad8-00fd7c9e5655 req-57318590-b1d3-4009-a6c4-c107b3227e01 service nova] Acquiring lock "refresh_cache-68e791b4-61db-4b6b-a30a-ccf5d6657643" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.959205] env[62183]: DEBUG oslo_concurrency.lockutils [req-57717834-aeef-4980-8ad8-00fd7c9e5655 req-57318590-b1d3-4009-a6c4-c107b3227e01 service nova] Acquired lock "refresh_cache-68e791b4-61db-4b6b-a30a-ccf5d6657643" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.959411] env[62183]: DEBUG nova.network.neutron [req-57717834-aeef-4980-8ad8-00fd7c9e5655 req-57318590-b1d3-4009-a6c4-c107b3227e01 service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Refreshing network info cache for port 1da5742b-a895-4c96-88f1-c84646ab800c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.201030] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.201510] env[62183]: DEBUG nova.compute.manager [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.204275] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bd48bb-4a37-2880-d97a-781738dc42a4, 'name': SearchDatastore_Task, 'duration_secs': 0.010553} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.204528] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.849s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.206270] env[62183]: INFO nova.compute.claims [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.209404] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.209706] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] ac555ffc-ce4e-4650-97fd-c26a3246fe4b/ac555ffc-ce4e-4650-97fd-c26a3246fe4b.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 834.210307] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5713d877-8805-4daa-bd6d-767803b22844 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.218657] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 834.218657] env[62183]: value = "task-1386991" [ 834.218657] env[62183]: _type = "Task" [ 834.218657] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.228462] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1386991, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.273663] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386987, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.285619} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.274532] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.274831] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac377f8-d813-4cbd-b825-7b82d1a82fb4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.297261] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 7465df88-c68c-49d3-9a91-6fff0d06957b/7465df88-c68c-49d3-9a91-6fff0d06957b.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.297547] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3db18748-8c5d-42cd-ae89-3e352717899c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.317267] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 834.317267] env[62183]: value = "task-1386992" [ 834.317267] env[62183]: _type = "Task" [ 834.317267] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.325125] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386992, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.342137] env[62183]: DEBUG nova.network.neutron [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Successfully updated port: eb7f6201-df05-4a22-b815-57d19d54b5fa {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.354939] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386990, 'name': CreateVM_Task, 'duration_secs': 0.445514} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.355648] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.356376] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.356557] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.356886] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.357450] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29b158a4-cec2-4bda-a29f-539c66d9268d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.362133] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 834.362133] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f046e0-6bd9-a970-a1c2-8e5d6b82c666" [ 834.362133] env[62183]: _type = "Task" [ 834.362133] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.369813] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f046e0-6bd9-a970-a1c2-8e5d6b82c666, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.706794] env[62183]: DEBUG nova.compute.utils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.709103] env[62183]: DEBUG nova.compute.manager [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 834.709103] env[62183]: DEBUG nova.network.neutron [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 834.729099] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1386991, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.759541] env[62183]: DEBUG nova.policy [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7deeb5ccaca42e29f0cb145a17e1097', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37d4936e2dc34e9d9ed73455081ee4ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 834.826255] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386992, 'name': ReconfigVM_Task, 'duration_secs': 0.29979} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.826552] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 7465df88-c68c-49d3-9a91-6fff0d06957b/7465df88-c68c-49d3-9a91-6fff0d06957b.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.827186] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0257ddc-058a-4714-a6da-202a892759e0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.833221] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 834.833221] env[62183]: value = "task-1386993" [ 834.833221] env[62183]: _type = "Task" [ 834.833221] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.840581] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386993, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.844246] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquiring lock "refresh_cache-39683931-d0ff-4a5c-a4a2-792230ab0e3d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.844383] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquired lock "refresh_cache-39683931-d0ff-4a5c-a4a2-792230ab0e3d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.844531] env[62183]: DEBUG nova.network.neutron [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.873802] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f046e0-6bd9-a970-a1c2-8e5d6b82c666, 'name': SearchDatastore_Task, 'duration_secs': 0.009579} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.874396] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.874396] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.874565] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.874703] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.874877] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.875152] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a1e40e6-15a7-430d-addf-2f9c65d32985 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.882776] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.882946] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.883679] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4dccec42-030b-4f90-97ed-1f07277d0d60 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.888998] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 834.888998] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e1c693-e83a-1fc1-451c-2da1d86be19b" [ 834.888998] env[62183]: _type = "Task" [ 834.888998] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.899072] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e1c693-e83a-1fc1-451c-2da1d86be19b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.946422] env[62183]: DEBUG nova.network.neutron [req-57717834-aeef-4980-8ad8-00fd7c9e5655 req-57318590-b1d3-4009-a6c4-c107b3227e01 service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Updated VIF entry in instance network info cache for port 1da5742b-a895-4c96-88f1-c84646ab800c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.946794] env[62183]: DEBUG nova.network.neutron [req-57717834-aeef-4980-8ad8-00fd7c9e5655 req-57318590-b1d3-4009-a6c4-c107b3227e01 service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Updating instance_info_cache with network_info: [{"id": "1da5742b-a895-4c96-88f1-c84646ab800c", "address": "fa:16:3e:57:d8:e2", "network": {"id": "559cdc13-7bba-43ae-8ddf-6c03d4efebb8", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1367814136-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "072024d50c3b49d09a721a3e3b7fd856", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1da5742b-a8", "ovs_interfaceid": "1da5742b-a895-4c96-88f1-c84646ab800c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.129673] env[62183]: DEBUG nova.network.neutron [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Successfully created port: d8838c27-a431-4260-88a0-4664366b6499 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.212640] env[62183]: DEBUG nova.compute.manager [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.231428] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1386991, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.343239] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386993, 'name': Rename_Task, 'duration_secs': 0.143475} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.343620] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 835.343888] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f903a1d-fcab-49a6-94b6-f12718512673 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.353834] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 835.353834] env[62183]: value = "task-1386994" [ 835.353834] env[62183]: _type = "Task" [ 835.353834] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.362184] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386994, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.376023] env[62183]: DEBUG nova.network.neutron [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.405094] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e1c693-e83a-1fc1-451c-2da1d86be19b, 'name': SearchDatastore_Task, 'duration_secs': 0.007994} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.405866] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dc0ed83-cff1-4918-8362-96007611789d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.409217] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c1fa5a-b745-4c1d-9f31-214bfaa92bdb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.414110] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 835.414110] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5291eadc-2cad-6b42-f74d-4fb4064f584e" [ 835.414110] env[62183]: _type = "Task" [ 835.414110] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.419467] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caed3874-c563-41df-a5c6-7fff28eded47 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.429626] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5291eadc-2cad-6b42-f74d-4fb4064f584e, 'name': SearchDatastore_Task, 'duration_secs': 0.008335} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.454158] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.454443] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 68e791b4-61db-4b6b-a30a-ccf5d6657643/68e791b4-61db-4b6b-a30a-ccf5d6657643.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.454932] env[62183]: DEBUG oslo_concurrency.lockutils [req-57717834-aeef-4980-8ad8-00fd7c9e5655 req-57318590-b1d3-4009-a6c4-c107b3227e01 service nova] Releasing lock "refresh_cache-68e791b4-61db-4b6b-a30a-ccf5d6657643" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.457774] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e0608d9c-f455-46f6-b1ad-8204061ad684 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.460061] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7c0120-9c1d-4e6e-890d-9819e15faee3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.468576] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f77300-df7d-4099-a80a-1e60ec1cac31 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.472266] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 835.472266] env[62183]: value = "task-1386995" [ 835.472266] env[62183]: _type = "Task" [ 835.472266] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.483199] env[62183]: DEBUG nova.compute.provider_tree [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.488735] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1386995, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.531928] env[62183]: DEBUG nova.network.neutron [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Updating instance_info_cache with network_info: [{"id": "eb7f6201-df05-4a22-b815-57d19d54b5fa", "address": "fa:16:3e:81:d4:5c", "network": {"id": "3ddcf714-0b6c-4c1d-a0bf-7b6192934f7a", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-938175151-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36152664258e4b0fbe576e8fbf1423ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb7f6201-df", "ovs_interfaceid": "eb7f6201-df05-4a22-b815-57d19d54b5fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.732983] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1386991, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.864044] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386994, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.982600] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1386995, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.986629] env[62183]: DEBUG nova.scheduler.client.report [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 836.015773] env[62183]: DEBUG nova.compute.manager [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Received event network-vif-plugged-eb7f6201-df05-4a22-b815-57d19d54b5fa {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.015773] env[62183]: DEBUG oslo_concurrency.lockutils [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] Acquiring lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.015953] env[62183]: DEBUG oslo_concurrency.lockutils [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] Lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.016117] env[62183]: DEBUG oslo_concurrency.lockutils [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] Lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.016293] env[62183]: DEBUG nova.compute.manager [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] No waiting events found dispatching network-vif-plugged-eb7f6201-df05-4a22-b815-57d19d54b5fa {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.016457] env[62183]: WARNING nova.compute.manager [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Received unexpected event network-vif-plugged-eb7f6201-df05-4a22-b815-57d19d54b5fa for instance with vm_state building and task_state spawning. [ 836.016614] env[62183]: DEBUG nova.compute.manager [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Received event network-changed-eb7f6201-df05-4a22-b815-57d19d54b5fa {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.016779] env[62183]: DEBUG nova.compute.manager [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Refreshing instance network info cache due to event network-changed-eb7f6201-df05-4a22-b815-57d19d54b5fa. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.016927] env[62183]: DEBUG oslo_concurrency.lockutils [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] Acquiring lock "refresh_cache-39683931-d0ff-4a5c-a4a2-792230ab0e3d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.034615] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Releasing lock "refresh_cache-39683931-d0ff-4a5c-a4a2-792230ab0e3d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.034941] env[62183]: DEBUG nova.compute.manager [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Instance network_info: |[{"id": "eb7f6201-df05-4a22-b815-57d19d54b5fa", "address": "fa:16:3e:81:d4:5c", "network": {"id": "3ddcf714-0b6c-4c1d-a0bf-7b6192934f7a", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-938175151-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36152664258e4b0fbe576e8fbf1423ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb7f6201-df", "ovs_interfaceid": "eb7f6201-df05-4a22-b815-57d19d54b5fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 836.035247] env[62183]: DEBUG oslo_concurrency.lockutils [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] Acquired lock "refresh_cache-39683931-d0ff-4a5c-a4a2-792230ab0e3d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.035425] env[62183]: DEBUG nova.network.neutron [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Refreshing network info cache for port eb7f6201-df05-4a22-b815-57d19d54b5fa {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 836.036602] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:d4:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd298db54-f13d-4bf6-b6c2-755074b3047f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb7f6201-df05-4a22-b815-57d19d54b5fa', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.043972] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Creating folder: Project (36152664258e4b0fbe576e8fbf1423ec). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.044626] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-907fd273-ad59-42de-939b-7e752a54a09d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.055443] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Created folder: Project (36152664258e4b0fbe576e8fbf1423ec) in parent group-v294392. [ 836.055654] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Creating folder: Instances. Parent ref: group-v294430. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 836.057127] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00a2562b-773f-474f-88ca-688f41fb89fa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.065719] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Created folder: Instances in parent group-v294430. [ 836.065965] env[62183]: DEBUG oslo.service.loopingcall [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.066170] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.066371] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-194a555f-7306-4c16-a3f2-293d3dbf19ea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.086955] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.086955] env[62183]: value = "task-1386998" [ 836.086955] env[62183]: _type = "Task" [ 836.086955] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.095670] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386998, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.227388] env[62183]: DEBUG nova.compute.manager [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.237938] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1386991, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.514815} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.238293] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] ac555ffc-ce4e-4650-97fd-c26a3246fe4b/ac555ffc-ce4e-4650-97fd-c26a3246fe4b.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.238553] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.238841] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60853086-ad4f-497f-92f6-934fff9c6de5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.245670] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 836.245670] env[62183]: value = "task-1386999" [ 836.245670] env[62183]: _type = "Task" [ 836.245670] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.256826] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1386999, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.259254] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.259530] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.259701] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.259895] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.260066] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.260282] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.260429] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.260566] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.260728] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.260886] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.261077] env[62183]: DEBUG nova.virt.hardware [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.262050] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596e1743-d61d-4a70-bec5-2286e7f502e8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.269974] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb724827-e32d-46b3-8e26-ad3810e488d8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.363799] env[62183]: DEBUG oslo_vmware.api [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1386994, 'name': PowerOnVM_Task, 'duration_secs': 0.546306} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.364513] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 836.364513] env[62183]: INFO nova.compute.manager [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Took 9.37 seconds to spawn the instance on the hypervisor. [ 836.364513] env[62183]: DEBUG nova.compute.manager [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.365227] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e2aea4-040a-4287-915a-0fc8df92163c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.482799] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1386995, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.491888] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.287s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.492194] env[62183]: DEBUG nova.compute.manager [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.495033] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.352s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.496275] env[62183]: INFO nova.compute.claims [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.559692] env[62183]: DEBUG nova.compute.manager [req-d6bef496-a182-4dd5-b664-8a63b8449b97 req-c020ad1d-7f95-4554-9fe3-314bd7678737 service nova] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Received event network-vif-plugged-d8838c27-a431-4260-88a0-4664366b6499 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.559916] env[62183]: DEBUG oslo_concurrency.lockutils [req-d6bef496-a182-4dd5-b664-8a63b8449b97 req-c020ad1d-7f95-4554-9fe3-314bd7678737 service nova] Acquiring lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.560137] env[62183]: DEBUG oslo_concurrency.lockutils [req-d6bef496-a182-4dd5-b664-8a63b8449b97 req-c020ad1d-7f95-4554-9fe3-314bd7678737 service nova] Lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.560310] env[62183]: DEBUG oslo_concurrency.lockutils [req-d6bef496-a182-4dd5-b664-8a63b8449b97 req-c020ad1d-7f95-4554-9fe3-314bd7678737 service nova] Lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.560478] env[62183]: DEBUG nova.compute.manager [req-d6bef496-a182-4dd5-b664-8a63b8449b97 req-c020ad1d-7f95-4554-9fe3-314bd7678737 service nova] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] No waiting events found dispatching network-vif-plugged-d8838c27-a431-4260-88a0-4664366b6499 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.560650] env[62183]: WARNING nova.compute.manager [req-d6bef496-a182-4dd5-b664-8a63b8449b97 req-c020ad1d-7f95-4554-9fe3-314bd7678737 service nova] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Received unexpected event network-vif-plugged-d8838c27-a431-4260-88a0-4664366b6499 for instance with vm_state building and task_state spawning. [ 836.601019] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1386998, 'name': CreateVM_Task, 'duration_secs': 0.369894} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.601019] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.601019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.601019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.601019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.601019] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88e3a66c-77bf-4dc9-95e8-c79c2998d2c9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.603383] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 836.603383] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525584a2-f237-9fbf-847f-f0be9f161c6b" [ 836.603383] env[62183]: _type = "Task" [ 836.603383] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.611194] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525584a2-f237-9fbf-847f-f0be9f161c6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.750441] env[62183]: DEBUG nova.network.neutron [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Updated VIF entry in instance network info cache for port eb7f6201-df05-4a22-b815-57d19d54b5fa. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.750845] env[62183]: DEBUG nova.network.neutron [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Updating instance_info_cache with network_info: [{"id": "eb7f6201-df05-4a22-b815-57d19d54b5fa", "address": "fa:16:3e:81:d4:5c", "network": {"id": "3ddcf714-0b6c-4c1d-a0bf-7b6192934f7a", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-938175151-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36152664258e4b0fbe576e8fbf1423ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d298db54-f13d-4bf6-b6c2-755074b3047f", "external-id": "nsx-vlan-transportzone-631", "segmentation_id": 631, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb7f6201-df", "ovs_interfaceid": "eb7f6201-df05-4a22-b815-57d19d54b5fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.758550] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1386999, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088461} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.758550] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.758550] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466d6bba-a1f8-4d36-a5a7-0d3c26608caf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.779379] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] ac555ffc-ce4e-4650-97fd-c26a3246fe4b/ac555ffc-ce4e-4650-97fd-c26a3246fe4b.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.779846] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a66612bf-4a94-445e-8e9f-c6daa341fa90 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.798533] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 836.798533] env[62183]: value = "task-1387000" [ 836.798533] env[62183]: _type = "Task" [ 836.798533] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.808428] env[62183]: DEBUG nova.network.neutron [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Successfully updated port: d8838c27-a431-4260-88a0-4664366b6499 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.809379] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387000, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.881937] env[62183]: INFO nova.compute.manager [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Took 34.27 seconds to build instance. [ 836.982926] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1386995, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.000379] env[62183]: DEBUG nova.compute.utils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.003894] env[62183]: DEBUG nova.compute.manager [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.004074] env[62183]: DEBUG nova.network.neutron [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.046855] env[62183]: DEBUG nova.policy [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b85b5dce9bf44da6afd11f100a121f03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '34581a9b5c1943eabc13bbb300a0f086', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.113118] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525584a2-f237-9fbf-847f-f0be9f161c6b, 'name': SearchDatastore_Task, 'duration_secs': 0.436753} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.113399] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.113723] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.113974] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.114121] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.114298] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.114559] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35a3a1ac-5f60-4762-a375-e6c56693bc5d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.122354] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.122644] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.123224] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5b94cab-817b-48d6-9a93-b30b0ed1d95c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.128275] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 837.128275] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]526f6fa0-b12a-b1b6-875d-c5d7e9be5d5f" [ 837.128275] env[62183]: _type = "Task" [ 837.128275] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.135802] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526f6fa0-b12a-b1b6-875d-c5d7e9be5d5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.253436] env[62183]: DEBUG oslo_concurrency.lockutils [req-ef0b77ec-c383-4915-b47f-904831b1a26e req-b803fab7-2e25-4a02-85ef-83668772af6f service nova] Releasing lock "refresh_cache-39683931-d0ff-4a5c-a4a2-792230ab0e3d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.308477] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387000, 'name': ReconfigVM_Task, 'duration_secs': 0.50411} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.309182] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Reconfigured VM instance instance-00000041 to attach disk [datastore1] ac555ffc-ce4e-4650-97fd-c26a3246fe4b/ac555ffc-ce4e-4650-97fd-c26a3246fe4b.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.309531] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3085d89-07b0-46ea-9ac1-e9a7d081d411 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.311553] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "refresh_cache-850d40d8-565a-49a2-a27f-3de2a8dc7e30" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.311767] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "refresh_cache-850d40d8-565a-49a2-a27f-3de2a8dc7e30" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.311951] env[62183]: DEBUG nova.network.neutron [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.318200] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 837.318200] env[62183]: value = "task-1387001" [ 837.318200] env[62183]: _type = "Task" [ 837.318200] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.329840] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387001, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.384163] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8948a1f-8757-4681-ae99-535cd31d8b30 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7465df88-c68c-49d3-9a91-6fff0d06957b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.272s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.453678] env[62183]: DEBUG nova.network.neutron [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Successfully created port: bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.483991] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1386995, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.591726} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.484279] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 68e791b4-61db-4b6b-a30a-ccf5d6657643/68e791b4-61db-4b6b-a30a-ccf5d6657643.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.484499] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.484768] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10e92ad9-3be7-41c6-b191-204b7cb7ca61 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.492140] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 837.492140] env[62183]: value = "task-1387002" [ 837.492140] env[62183]: _type = "Task" [ 837.492140] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.500038] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387002, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.505085] env[62183]: DEBUG nova.compute.manager [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.639271] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526f6fa0-b12a-b1b6-875d-c5d7e9be5d5f, 'name': SearchDatastore_Task, 'duration_secs': 0.00884} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.640176] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f6ffb03-df6c-4c5e-a427-a78c17c058cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.646011] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 837.646011] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]529a51bf-74e5-9bed-60e0-264bb8eb7818" [ 837.646011] env[62183]: _type = "Task" [ 837.646011] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.657232] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529a51bf-74e5-9bed-60e0-264bb8eb7818, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.736784] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b9c7f1-4d73-48a9-bfcc-ce0178845712 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.743822] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dddacc9e-c9ad-4ca2-99d2-b93c74f87e1d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.774161] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb142b8-0040-42b4-8d7d-2167d7aeea40 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.781445] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f4c49c-c2ad-4190-82bb-4c8941970207 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.795873] env[62183]: DEBUG nova.compute.provider_tree [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 837.828924] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387001, 'name': Rename_Task, 'duration_secs': 0.147135} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.829210] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.829456] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a89301d9-2135-46f4-87bc-c0d80275789c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.834819] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 837.834819] env[62183]: value = "task-1387003" [ 837.834819] env[62183]: _type = "Task" [ 837.834819] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.842122] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387003, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.867056] env[62183]: DEBUG nova.network.neutron [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.886484] env[62183]: DEBUG nova.compute.manager [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.006023] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387002, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.484689} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.006023] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.006023] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71dc89d-ab83-4b04-8484-26f85e4f34b7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.038891] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 68e791b4-61db-4b6b-a30a-ccf5d6657643/68e791b4-61db-4b6b-a30a-ccf5d6657643.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.042603] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dc2b002-043c-439d-b1d8-ca1103751afe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.069482] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 838.069482] env[62183]: value = "task-1387004" [ 838.069482] env[62183]: _type = "Task" [ 838.069482] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.074563] env[62183]: DEBUG nova.network.neutron [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Updating instance_info_cache with network_info: [{"id": "d8838c27-a431-4260-88a0-4664366b6499", "address": "fa:16:3e:25:64:6b", "network": {"id": "bef4069c-60bb-4ca3-b068-b7d33e5c9d90", "bridge": "br-int", "label": "tempest-ImagesTestJSON-798269247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37d4936e2dc34e9d9ed73455081ee4ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8838c27-a4", "ovs_interfaceid": "d8838c27-a431-4260-88a0-4664366b6499", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.080602] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387004, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.157297] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529a51bf-74e5-9bed-60e0-264bb8eb7818, 'name': SearchDatastore_Task, 'duration_secs': 0.317294} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.157297] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.157565] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 39683931-d0ff-4a5c-a4a2-792230ab0e3d/39683931-d0ff-4a5c-a4a2-792230ab0e3d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.158646] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fbaa379-e8f6-4418-b0ca-c27e8bdd8c9c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.163896] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 838.163896] env[62183]: value = "task-1387005" [ 838.163896] env[62183]: _type = "Task" [ 838.163896] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.172473] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387005, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.329978] env[62183]: ERROR nova.scheduler.client.report [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [req-08590203-983d-4199-9e4b-c21272e58bfa] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-08590203-983d-4199-9e4b-c21272e58bfa"}]} [ 838.344751] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387003, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.352983] env[62183]: DEBUG nova.scheduler.client.report [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 838.369141] env[62183]: DEBUG nova.scheduler.client.report [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 838.369452] env[62183]: DEBUG nova.compute.provider_tree [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 838.396189] env[62183]: DEBUG nova.scheduler.client.report [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 838.413449] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.419347] env[62183]: DEBUG nova.scheduler.client.report [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 838.545731] env[62183]: DEBUG nova.compute.manager [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.579593] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.579883] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.580098] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.580326] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.580527] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.580696] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.580905] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.581076] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.581249] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.581442] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.581648] env[62183]: DEBUG nova.virt.hardware [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.582266] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "refresh_cache-850d40d8-565a-49a2-a27f-3de2a8dc7e30" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.582574] env[62183]: DEBUG nova.compute.manager [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Instance network_info: |[{"id": "d8838c27-a431-4260-88a0-4664366b6499", "address": "fa:16:3e:25:64:6b", "network": {"id": "bef4069c-60bb-4ca3-b068-b7d33e5c9d90", "bridge": "br-int", "label": "tempest-ImagesTestJSON-798269247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37d4936e2dc34e9d9ed73455081ee4ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8838c27-a4", "ovs_interfaceid": "d8838c27-a431-4260-88a0-4664366b6499", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 838.583470] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd21663-f94e-4b2b-bf35-db989d3adb0b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.589196] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:64:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11032cc2-b275-48d2-9c40-9455ea7d49e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd8838c27-a431-4260-88a0-4664366b6499', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.597422] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Creating folder: Project (37d4936e2dc34e9d9ed73455081ee4ab). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.599655] env[62183]: DEBUG nova.compute.manager [req-4a6effe5-826c-43d2-bd65-8d127f4ab52b req-0619d5d1-38c2-4a07-a138-51d01bc5ca5f service nova] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Received event network-changed-d8838c27-a431-4260-88a0-4664366b6499 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.599655] env[62183]: DEBUG nova.compute.manager [req-4a6effe5-826c-43d2-bd65-8d127f4ab52b req-0619d5d1-38c2-4a07-a138-51d01bc5ca5f service nova] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Refreshing instance network info cache due to event network-changed-d8838c27-a431-4260-88a0-4664366b6499. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.599945] env[62183]: DEBUG oslo_concurrency.lockutils [req-4a6effe5-826c-43d2-bd65-8d127f4ab52b req-0619d5d1-38c2-4a07-a138-51d01bc5ca5f service nova] Acquiring lock "refresh_cache-850d40d8-565a-49a2-a27f-3de2a8dc7e30" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.599945] env[62183]: DEBUG oslo_concurrency.lockutils [req-4a6effe5-826c-43d2-bd65-8d127f4ab52b req-0619d5d1-38c2-4a07-a138-51d01bc5ca5f service nova] Acquired lock "refresh_cache-850d40d8-565a-49a2-a27f-3de2a8dc7e30" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.600139] env[62183]: DEBUG nova.network.neutron [req-4a6effe5-826c-43d2-bd65-8d127f4ab52b req-0619d5d1-38c2-4a07-a138-51d01bc5ca5f service nova] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Refreshing network info cache for port d8838c27-a431-4260-88a0-4664366b6499 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.607926] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f6c4a76-dd1a-4a75-9bbd-646e5d65912e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.611807] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387004, 'name': ReconfigVM_Task, 'duration_secs': 0.296396} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.617681] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 68e791b4-61db-4b6b-a30a-ccf5d6657643/68e791b4-61db-4b6b-a30a-ccf5d6657643.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.619270] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8d69f9e-845f-4eb7-843b-53a2717ee32d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.622736] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51bd7c09-41bf-445a-bf3f-1246ac1ccae6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.632573] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Created folder: Project (37d4936e2dc34e9d9ed73455081ee4ab) in parent group-v294392. [ 838.633551] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Creating folder: Instances. Parent ref: group-v294433. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.635050] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e58ebf16-3237-4fa0-952c-9a2ca1d79629 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.641619] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 838.641619] env[62183]: value = "task-1387007" [ 838.641619] env[62183]: _type = "Task" [ 838.641619] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.654987] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Created folder: Instances in parent group-v294433. [ 838.655281] env[62183]: DEBUG oslo.service.loopingcall [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.658598] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 838.659353] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a01fd7b-0e27-496c-8373-427faa8a5dc8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.682854] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387007, 'name': Rename_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.697141] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.697141] env[62183]: value = "task-1387009" [ 838.697141] env[62183]: _type = "Task" [ 838.697141] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.697440] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387005, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507222} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.698574] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 39683931-d0ff-4a5c-a4a2-792230ab0e3d/39683931-d0ff-4a5c-a4a2-792230ab0e3d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.698802] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.699548] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4531843-c0f5-406b-a008-f48a900c9082 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.705588] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-accef19f-e811-4552-98c8-74b8eccc76e7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.714015] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd63171-45fd-4ce6-9041-d2018ee1572f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.719377] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 838.719377] env[62183]: value = "task-1387010" [ 838.719377] env[62183]: _type = "Task" [ 838.719377] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.752573] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833f5664-b96e-48d7-950a-fe811dc235b9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.758510] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387010, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.763948] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ed71f0-11c3-4354-9063-4844c2f853b7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.779077] env[62183]: DEBUG nova.compute.provider_tree [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 838.787414] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.787835] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.845228] env[62183]: DEBUG oslo_vmware.api [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387003, 'name': PowerOnVM_Task, 'duration_secs': 0.523516} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.845509] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.845858] env[62183]: INFO nova.compute.manager [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Took 9.58 seconds to spawn the instance on the hypervisor. [ 838.845934] env[62183]: DEBUG nova.compute.manager [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.846752] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f1bfd9-3232-44d8-bc3b-c6b88cefd695 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.971792] env[62183]: DEBUG nova.network.neutron [req-4a6effe5-826c-43d2-bd65-8d127f4ab52b req-0619d5d1-38c2-4a07-a138-51d01bc5ca5f service nova] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Updated VIF entry in instance network info cache for port d8838c27-a431-4260-88a0-4664366b6499. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.972262] env[62183]: DEBUG nova.network.neutron [req-4a6effe5-826c-43d2-bd65-8d127f4ab52b req-0619d5d1-38c2-4a07-a138-51d01bc5ca5f service nova] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Updating instance_info_cache with network_info: [{"id": "d8838c27-a431-4260-88a0-4664366b6499", "address": "fa:16:3e:25:64:6b", "network": {"id": "bef4069c-60bb-4ca3-b068-b7d33e5c9d90", "bridge": "br-int", "label": "tempest-ImagesTestJSON-798269247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37d4936e2dc34e9d9ed73455081ee4ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8838c27-a4", "ovs_interfaceid": "d8838c27-a431-4260-88a0-4664366b6499", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.091948] env[62183]: DEBUG nova.network.neutron [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Successfully updated port: bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.151695] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387007, 'name': Rename_Task, 'duration_secs': 0.151451} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.151959] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 839.152211] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f57e2cfa-82ec-4a12-bcb2-5ddc37b3d4c4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.158802] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 839.158802] env[62183]: value = "task-1387011" [ 839.158802] env[62183]: _type = "Task" [ 839.158802] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.166327] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.207758] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387009, 'name': CreateVM_Task, 'duration_secs': 0.345893} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.207922] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 839.208592] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.208757] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.209083] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 839.209325] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54b04441-69e7-4de4-8a18-a313ec29b056 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.213683] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 839.213683] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52743ff6-a552-e280-3b27-f7aca1b34e8c" [ 839.213683] env[62183]: _type = "Task" [ 839.213683] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.221355] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52743ff6-a552-e280-3b27-f7aca1b34e8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.228395] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387010, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065557} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.228645] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.229392] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f7073b-1fdd-4b1e-9647-5d7f25a2330a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.251933] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 39683931-d0ff-4a5c-a4a2-792230ab0e3d/39683931-d0ff-4a5c-a4a2-792230ab0e3d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.252235] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6cd8f5bd-d0ff-406d-96ee-83cfd3f1f2d2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.271536] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 839.271536] env[62183]: value = "task-1387012" [ 839.271536] env[62183]: _type = "Task" [ 839.271536] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.279305] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387012, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.319240] env[62183]: DEBUG nova.scheduler.client.report [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 83 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 839.319522] env[62183]: DEBUG nova.compute.provider_tree [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 83 to 84 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 839.319758] env[62183]: DEBUG nova.compute.provider_tree [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 839.367907] env[62183]: INFO nova.compute.manager [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Took 35.07 seconds to build instance. [ 839.475305] env[62183]: DEBUG oslo_concurrency.lockutils [req-4a6effe5-826c-43d2-bd65-8d127f4ab52b req-0619d5d1-38c2-4a07-a138-51d01bc5ca5f service nova] Releasing lock "refresh_cache-850d40d8-565a-49a2-a27f-3de2a8dc7e30" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.598326] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.598326] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.598497] env[62183]: DEBUG nova.network.neutron [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.669419] env[62183]: DEBUG oslo_vmware.api [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387011, 'name': PowerOnVM_Task, 'duration_secs': 0.495938} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.670030] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.670030] env[62183]: INFO nova.compute.manager [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Took 8.08 seconds to spawn the instance on the hypervisor. [ 839.670227] env[62183]: DEBUG nova.compute.manager [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.670934] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb74df4-e8a0-4afb-8127-a3915b8151c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.725159] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52743ff6-a552-e280-3b27-f7aca1b34e8c, 'name': SearchDatastore_Task, 'duration_secs': 0.132194} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.725458] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.725684] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.726153] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.726153] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.726278] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.726490] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-336eb777-bb68-4b8f-9c3b-d9bd6662b330 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.740074] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.740262] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.740977] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa540da5-2e9d-4e34-aed3-9825a333fbc7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.747903] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 839.747903] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a8a9c8-f5a8-9ce0-ac5f-c99c18167bff" [ 839.747903] env[62183]: _type = "Task" [ 839.747903] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.755763] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a8a9c8-f5a8-9ce0-ac5f-c99c18167bff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.781049] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387012, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.824950] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.330s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.825486] env[62183]: DEBUG nova.compute.manager [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.828036] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.954s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.869922] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e8bf4b7-757a-439d-824f-23620a26f712 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.220s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.139674] env[62183]: DEBUG nova.network.neutron [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.197471] env[62183]: INFO nova.compute.manager [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Took 34.03 seconds to build instance. [ 840.261497] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a8a9c8-f5a8-9ce0-ac5f-c99c18167bff, 'name': SearchDatastore_Task, 'duration_secs': 0.030794} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.262830] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3615a01-167c-4ec8-8b3d-8cad103ebf01 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.268521] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 840.268521] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]520a8d78-f793-3705-c655-b0928702d27d" [ 840.268521] env[62183]: _type = "Task" [ 840.268521] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.279059] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]520a8d78-f793-3705-c655-b0928702d27d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.283433] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387012, 'name': ReconfigVM_Task, 'duration_secs': 0.789561} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.283723] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 39683931-d0ff-4a5c-a4a2-792230ab0e3d/39683931-d0ff-4a5c-a4a2-792230ab0e3d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.284341] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a895379-1bff-4379-ac0e-c05a16009e83 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.290675] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 840.290675] env[62183]: value = "task-1387013" [ 840.290675] env[62183]: _type = "Task" [ 840.290675] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.301841] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387013, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.332744] env[62183]: DEBUG nova.compute.utils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.334549] env[62183]: DEBUG nova.compute.manager [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Not allocating networking since 'none' was specified. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 840.372500] env[62183]: DEBUG nova.compute.manager [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 840.382684] env[62183]: DEBUG nova.network.neutron [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.562440] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046e9fb4-b0d2-4bac-9062-5b170ede0748 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.570399] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9cb4bfb-372f-4ca4-a181-fcb9c7e2a92e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.602282] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520416a2-2f33-4db3-ad66-9c596fbcf07e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.609339] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f333dd46-74ca-46d9-b0c8-da41f4f337fa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.623267] env[62183]: DEBUG nova.compute.provider_tree [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.647108] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquiring lock "68e791b4-61db-4b6b-a30a-ccf5d6657643" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.654674] env[62183]: DEBUG nova.compute.manager [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Received event network-vif-plugged-bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.654899] env[62183]: DEBUG oslo_concurrency.lockutils [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] Acquiring lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.655122] env[62183]: DEBUG oslo_concurrency.lockutils [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.655329] env[62183]: DEBUG oslo_concurrency.lockutils [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.655536] env[62183]: DEBUG nova.compute.manager [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] No waiting events found dispatching network-vif-plugged-bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.655701] env[62183]: WARNING nova.compute.manager [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Received unexpected event network-vif-plugged-bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e for instance with vm_state building and task_state spawning. [ 840.655871] env[62183]: DEBUG nova.compute.manager [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Received event network-changed-bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.656081] env[62183]: DEBUG nova.compute.manager [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Refreshing instance network info cache due to event network-changed-bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.656201] env[62183]: DEBUG oslo_concurrency.lockutils [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] Acquiring lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.699256] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e48364d3-1659-4668-980c-b8c8c23fe2d4 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "68e791b4-61db-4b6b-a30a-ccf5d6657643" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.965s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.700459] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "68e791b4-61db-4b6b-a30a-ccf5d6657643" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.054s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.700680] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquiring lock "68e791b4-61db-4b6b-a30a-ccf5d6657643-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.700882] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "68e791b4-61db-4b6b-a30a-ccf5d6657643-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.701061] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "68e791b4-61db-4b6b-a30a-ccf5d6657643-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.704296] env[62183]: INFO nova.compute.manager [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Terminating instance [ 840.705923] env[62183]: DEBUG nova.compute.manager [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.706132] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.706946] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324f9f82-094f-4c34-8eb0-52011d78dbbe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.714433] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.715169] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f20ca6e6-6976-4300-9cac-7d244036c670 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.721865] env[62183]: DEBUG oslo_vmware.api [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 840.721865] env[62183]: value = "task-1387014" [ 840.721865] env[62183]: _type = "Task" [ 840.721865] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.729417] env[62183]: DEBUG oslo_vmware.api [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.779202] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]520a8d78-f793-3705-c655-b0928702d27d, 'name': SearchDatastore_Task, 'duration_secs': 0.019453} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.779548] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.779861] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 850d40d8-565a-49a2-a27f-3de2a8dc7e30/850d40d8-565a-49a2-a27f-3de2a8dc7e30.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.780170] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-315bdf16-617d-4750-8b97-5f35c0e7863f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.787959] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 840.787959] env[62183]: value = "task-1387015" [ 840.787959] env[62183]: _type = "Task" [ 840.787959] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.799650] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387015, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.803026] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387013, 'name': Rename_Task, 'duration_secs': 0.164945} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.803138] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.803385] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94bdf864-10bd-40ea-ab04-87f7cf7bc1ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.809895] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 840.809895] env[62183]: value = "task-1387016" [ 840.809895] env[62183]: _type = "Task" [ 840.809895] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.819129] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387016, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.838064] env[62183]: DEBUG nova.compute.manager [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.889014] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.889014] env[62183]: DEBUG nova.compute.manager [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Instance network_info: |[{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 840.889540] env[62183]: DEBUG oslo_concurrency.lockutils [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] Acquired lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.889663] env[62183]: DEBUG nova.network.neutron [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Refreshing network info cache for port bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.891246] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:83:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8800a981-a89e-42e4-8be9-cace419ba9cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.904043] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Creating folder: Project (34581a9b5c1943eabc13bbb300a0f086). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.904814] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.908024] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3eaee1c-2da1-4e84-a9e4-d1a22eda30fd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.920664] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Created folder: Project (34581a9b5c1943eabc13bbb300a0f086) in parent group-v294392. [ 840.920890] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Creating folder: Instances. Parent ref: group-v294436. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 840.921174] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b9e56739-37f1-4214-85f2-e0bc3dfff2b7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.931257] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Created folder: Instances in parent group-v294436. [ 840.931536] env[62183]: DEBUG oslo.service.loopingcall [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.931789] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 840.931985] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b52e8cd3-5248-4bc7-8be9-80c1c2beb85c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.954061] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.954061] env[62183]: value = "task-1387019" [ 840.954061] env[62183]: _type = "Task" [ 840.954061] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.962598] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387019, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.127210] env[62183]: DEBUG nova.scheduler.client.report [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 841.164762] env[62183]: DEBUG nova.network.neutron [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updated VIF entry in instance network info cache for port bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 841.165186] env[62183]: DEBUG nova.network.neutron [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.203701] env[62183]: DEBUG nova.compute.manager [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 841.231558] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.231906] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.237795] env[62183]: DEBUG oslo_vmware.api [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387014, 'name': PowerOffVM_Task, 'duration_secs': 0.177907} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.238549] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 841.239113] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 841.239172] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-378dd8d6-1c61-4c78-a5ec-dcb2d8c66e6d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.299244] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387015, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.321196] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387016, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.338895] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.339089] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.339374] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Deleting the datastore file [datastore1] 68e791b4-61db-4b6b-a30a-ccf5d6657643 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.340077] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca273482-7f7c-4818-9e52-7e237dcd3d7b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.351958] env[62183]: DEBUG oslo_vmware.api [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for the task: (returnval){ [ 841.351958] env[62183]: value = "task-1387021" [ 841.351958] env[62183]: _type = "Task" [ 841.351958] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.361296] env[62183]: DEBUG oslo_vmware.api [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387021, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.464446] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387019, 'name': CreateVM_Task, 'duration_secs': 0.370374} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.464689] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 841.465536] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.465766] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.466225] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 841.467033] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21bcf129-e510-4b72-b841-042f86448824 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.472114] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 841.472114] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5261a59a-0014-6566-1ef0-78637abffd96" [ 841.472114] env[62183]: _type = "Task" [ 841.472114] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.480685] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5261a59a-0014-6566-1ef0-78637abffd96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.633054] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.805s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.633678] env[62183]: ERROR nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port abf57cfb-a5e5-4710-8eeb-bec026db61f5, please check neutron logs for more information. [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Traceback (most recent call last): [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self.driver.spawn(context, instance, image_meta, [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self._vmops.spawn(context, instance, image_meta, injected_files, [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] vm_ref = self.build_virtual_machine(instance, [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] vif_infos = vmwarevif.get_vif_info(self._session, [ 841.633678] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] for vif in network_info: [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] return self._sync_wrapper(fn, *args, **kwargs) [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self.wait() [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self[:] = self._gt.wait() [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] return self._exit_event.wait() [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] result = hub.switch() [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 841.634051] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] return self.greenlet.switch() [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] result = function(*args, **kwargs) [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] return func(*args, **kwargs) [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] raise e [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] nwinfo = self.network_api.allocate_for_instance( [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] created_port_ids = self._update_ports_for_instance( [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] with excutils.save_and_reraise_exception(): [ 841.634366] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] self.force_reraise() [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] raise self.value [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] updated_port = self._update_port( [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] _ensure_no_port_binding_failure(port) [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] raise exception.PortBindingFailed(port_id=port['id']) [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] nova.exception.PortBindingFailed: Binding failed for port abf57cfb-a5e5-4710-8eeb-bec026db61f5, please check neutron logs for more information. [ 841.634744] env[62183]: ERROR nova.compute.manager [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] [ 841.635070] env[62183]: DEBUG nova.compute.utils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Binding failed for port abf57cfb-a5e5-4710-8eeb-bec026db61f5, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 841.635713] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.434s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.638869] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Build of instance 0dd90b88-ed92-4331-98c5-c7481ac1ae15 was re-scheduled: Binding failed for port abf57cfb-a5e5-4710-8eeb-bec026db61f5, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 841.639029] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 841.639185] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Acquiring lock "refresh_cache-0dd90b88-ed92-4331-98c5-c7481ac1ae15" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.639331] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Acquired lock "refresh_cache-0dd90b88-ed92-4331-98c5-c7481ac1ae15" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.639490] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.668976] env[62183]: DEBUG oslo_concurrency.lockutils [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] Releasing lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.669470] env[62183]: DEBUG nova.compute.manager [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.670869] env[62183]: DEBUG nova.compute.manager [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing instance network info cache due to event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 841.670869] env[62183]: DEBUG oslo_concurrency.lockutils [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] Acquiring lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.670869] env[62183]: DEBUG oslo_concurrency.lockutils [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] Acquired lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.670869] env[62183]: DEBUG nova.network.neutron [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 841.729052] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.798084] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387015, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534853} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.798341] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 850d40d8-565a-49a2-a27f-3de2a8dc7e30/850d40d8-565a-49a2-a27f-3de2a8dc7e30.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 841.798599] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.798782] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90afdcc5-d707-4ae7-8a8a-206dd821d8a7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.805696] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 841.805696] env[62183]: value = "task-1387022" [ 841.805696] env[62183]: _type = "Task" [ 841.805696] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.814691] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387022, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.823233] env[62183]: DEBUG oslo_vmware.api [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387016, 'name': PowerOnVM_Task, 'duration_secs': 0.566119} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.823529] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.823778] env[62183]: INFO nova.compute.manager [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Took 7.94 seconds to spawn the instance on the hypervisor. [ 841.823991] env[62183]: DEBUG nova.compute.manager [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.824823] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b8bee5-fb0d-4750-92bb-00a5be956365 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.848520] env[62183]: DEBUG nova.compute.manager [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.861438] env[62183]: DEBUG oslo_vmware.api [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387021, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.874708] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.874964] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.875136] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.875317] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.875464] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.875607] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.875811] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.875967] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.876162] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.876325] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.876493] env[62183]: DEBUG nova.virt.hardware [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.877499] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ae7185-f863-49c6-9292-2a6d40ff5c48 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.884793] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e99e94-c588-4653-8d6d-910aa778d0b2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.897870] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.903604] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Creating folder: Project (26e35e92331e48eabd2911f4d8b431aa). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 841.903684] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ba0dc63-afbe-4a9d-a804-9ee91ce83474 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.912759] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Created folder: Project (26e35e92331e48eabd2911f4d8b431aa) in parent group-v294392. [ 841.912943] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Creating folder: Instances. Parent ref: group-v294439. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 841.913170] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69c2295f-aebe-4420-8df0-a3a7129a08b8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.921487] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Created folder: Instances in parent group-v294439. [ 841.921733] env[62183]: DEBUG oslo.service.loopingcall [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.921917] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 841.922114] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96537325-7572-460a-b5c6-b5e1338bc5d3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.938022] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.938022] env[62183]: value = "task-1387025" [ 841.938022] env[62183]: _type = "Task" [ 841.938022] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.947733] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387025, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.981956] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5261a59a-0014-6566-1ef0-78637abffd96, 'name': SearchDatastore_Task, 'duration_secs': 0.039266} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.982305] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.982538] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.982791] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.982942] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.983191] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.983406] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a49b2de-5a9a-44a9-8d45-67be0d0f6778 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.992358] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.992540] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.993275] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1a71e07-6614-4459-ae28-92f5b326a229 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.998407] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 841.998407] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52096cac-dfdf-56b2-d350-00d5bc7cf584" [ 841.998407] env[62183]: _type = "Task" [ 841.998407] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.006179] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52096cac-dfdf-56b2-d350-00d5bc7cf584, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.164203] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.249484] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.316529] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387022, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063859} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.316789] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 842.317609] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee29f64-873e-44e2-a057-b4eb36ed05aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.341567] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 850d40d8-565a-49a2-a27f-3de2a8dc7e30/850d40d8-565a-49a2-a27f-3de2a8dc7e30.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.349808] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-940aed23-e282-4df6-b107-010ac16b86c5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.367201] env[62183]: INFO nova.compute.manager [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Took 32.16 seconds to build instance. [ 842.379766] env[62183]: DEBUG oslo_vmware.api [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Task: {'id': task-1387021, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.646081} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.381215] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.381404] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 842.381587] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 842.381768] env[62183]: INFO nova.compute.manager [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Took 1.68 seconds to destroy the instance on the hypervisor. [ 842.382009] env[62183]: DEBUG oslo.service.loopingcall [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.382579] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 842.382579] env[62183]: value = "task-1387026" [ 842.382579] env[62183]: _type = "Task" [ 842.382579] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.383419] env[62183]: DEBUG nova.compute.manager [-] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.383520] env[62183]: DEBUG nova.network.neutron [-] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 842.391032] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffaca6c-7b35-4139-854f-74e061877abc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.400411] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387026, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.401515] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872caa03-4259-49fe-b049-b737cfea04a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.439059] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763e4fd0-4997-4f9b-9e85-2477fda36c73 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.451957] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387025, 'name': CreateVM_Task, 'duration_secs': 0.302536} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.456184] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 842.456660] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.456815] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.457147] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.458654] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cbeb324-d219-43b6-8352-eae9f31a0861 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.462381] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ec6fe2f-9c7d-4da2-bd2e-02c9b777cf01 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.467803] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 842.467803] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52da6f6f-957a-431e-b21b-18d829fde967" [ 842.467803] env[62183]: _type = "Task" [ 842.467803] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.476610] env[62183]: DEBUG nova.compute.provider_tree [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.486480] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52da6f6f-957a-431e-b21b-18d829fde967, 'name': SearchDatastore_Task, 'duration_secs': 0.014937} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.487595] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.487680] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.487904] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.494835] env[62183]: DEBUG nova.network.neutron [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updated VIF entry in instance network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 842.495230] env[62183]: DEBUG nova.network.neutron [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.508593] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52096cac-dfdf-56b2-d350-00d5bc7cf584, 'name': SearchDatastore_Task, 'duration_secs': 0.031404} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.509374] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e20ffd6f-967c-41af-b658-9a81937cb9c0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.514798] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 842.514798] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52c2f3ee-0047-a76f-ff6b-98eb0d023eff" [ 842.514798] env[62183]: _type = "Task" [ 842.514798] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.522499] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c2f3ee-0047-a76f-ff6b-98eb0d023eff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.752541] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Releasing lock "refresh_cache-0dd90b88-ed92-4331-98c5-c7481ac1ae15" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.753995] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 842.754242] env[62183]: DEBUG nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.754414] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 842.773435] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.872892] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d9747f7d-f9e4-4164-838d-d8c2d0f1dc8e tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.948s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.877238] env[62183]: DEBUG nova.compute.manager [req-d6036103-9d60-4d5c-84f0-cd60b6ae2609 req-f0fef532-507b-440e-9e49-e40202df64dc service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Received event network-vif-deleted-1da5742b-a895-4c96-88f1-c84646ab800c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.877238] env[62183]: INFO nova.compute.manager [req-d6036103-9d60-4d5c-84f0-cd60b6ae2609 req-f0fef532-507b-440e-9e49-e40202df64dc service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Neutron deleted interface 1da5742b-a895-4c96-88f1-c84646ab800c; detaching it from the instance and deleting it from the info cache [ 842.877238] env[62183]: DEBUG nova.network.neutron [req-d6036103-9d60-4d5c-84f0-cd60b6ae2609 req-f0fef532-507b-440e-9e49-e40202df64dc service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.894706] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387026, 'name': ReconfigVM_Task, 'duration_secs': 0.291301} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.894945] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 850d40d8-565a-49a2-a27f-3de2a8dc7e30/850d40d8-565a-49a2-a27f-3de2a8dc7e30.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.895563] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8f05b6f-4d8c-4e6e-94fb-170234a5eacd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.902032] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 842.902032] env[62183]: value = "task-1387027" [ 842.902032] env[62183]: _type = "Task" [ 842.902032] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.909333] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387027, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.982410] env[62183]: DEBUG nova.scheduler.client.report [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 842.997999] env[62183]: DEBUG oslo_concurrency.lockutils [req-92f3ffcc-9b4c-4d64-8d98-f32be3929589 req-475ec728-79d7-48d7-a24c-2e33a68af590 service nova] Releasing lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.026618] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c2f3ee-0047-a76f-ff6b-98eb0d023eff, 'name': SearchDatastore_Task, 'duration_secs': 0.008871} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.026618] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.026618] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a/a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 843.026618] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.026917] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.026954] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-927ba529-525b-41ad-b3db-61888570beab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.029257] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66210ac1-b82c-4850-a575-c0e7fcee2f38 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.036210] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 843.036210] env[62183]: value = "task-1387028" [ 843.036210] env[62183]: _type = "Task" [ 843.036210] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.040111] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.040286] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.041266] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9407d562-d22c-4055-ad9e-de5f613745c8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.048904] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387028, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.051651] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 843.051651] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5288fdce-8079-0023-d1fe-c50818852ef7" [ 843.051651] env[62183]: _type = "Task" [ 843.051651] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.058783] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5288fdce-8079-0023-d1fe-c50818852ef7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.211413] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f958ff6b-3a5d-4e67-9920-58b6ceae5a9f tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquiring lock "interface-39683931-d0ff-4a5c-a4a2-792230ab0e3d-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.211793] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f958ff6b-3a5d-4e67-9920-58b6ceae5a9f tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "interface-39683931-d0ff-4a5c-a4a2-792230ab0e3d-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.212240] env[62183]: DEBUG nova.objects.instance [None req-f958ff6b-3a5d-4e67-9920-58b6ceae5a9f tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lazy-loading 'flavor' on Instance uuid 39683931-d0ff-4a5c-a4a2-792230ab0e3d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.218499] env[62183]: DEBUG nova.network.neutron [-] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.276193] env[62183]: DEBUG nova.network.neutron [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.375417] env[62183]: DEBUG nova.compute.manager [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.379390] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-940902d8-75c6-4d9d-b505-0b9bd1bb98fd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.389009] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd22dea-966d-4f9a-9606-f5b48e75c770 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.411379] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387027, 'name': Rename_Task, 'duration_secs': 0.129451} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.422401] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 843.426848] env[62183]: DEBUG nova.compute.manager [req-d6036103-9d60-4d5c-84f0-cd60b6ae2609 req-f0fef532-507b-440e-9e49-e40202df64dc service nova] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Detach interface failed, port_id=1da5742b-a895-4c96-88f1-c84646ab800c, reason: Instance 68e791b4-61db-4b6b-a30a-ccf5d6657643 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 843.427351] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d716927c-1f9c-4359-bc1c-76ade2e9e909 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.441054] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 843.441054] env[62183]: value = "task-1387029" [ 843.441054] env[62183]: _type = "Task" [ 843.441054] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.456273] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387029, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.487412] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.851s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.488123] env[62183]: ERROR nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8af4e5da-0239-4399-bc9e-79525f73ce64, please check neutron logs for more information. [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Traceback (most recent call last): [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self.driver.spawn(context, instance, image_meta, [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] vm_ref = self.build_virtual_machine(instance, [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] vif_infos = vmwarevif.get_vif_info(self._session, [ 843.488123] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] for vif in network_info: [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] return self._sync_wrapper(fn, *args, **kwargs) [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self.wait() [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self[:] = self._gt.wait() [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] return self._exit_event.wait() [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] current.throw(*self._exc) [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 843.488486] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] result = function(*args, **kwargs) [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] return func(*args, **kwargs) [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] raise e [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] nwinfo = self.network_api.allocate_for_instance( [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] created_port_ids = self._update_ports_for_instance( [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] with excutils.save_and_reraise_exception(): [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] self.force_reraise() [ 843.488790] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 843.489161] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] raise self.value [ 843.489161] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 843.489161] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] updated_port = self._update_port( [ 843.489161] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 843.489161] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] _ensure_no_port_binding_failure(port) [ 843.489161] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 843.489161] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] raise exception.PortBindingFailed(port_id=port['id']) [ 843.489161] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] nova.exception.PortBindingFailed: Binding failed for port 8af4e5da-0239-4399-bc9e-79525f73ce64, please check neutron logs for more information. [ 843.489161] env[62183]: ERROR nova.compute.manager [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] [ 843.489161] env[62183]: DEBUG nova.compute.utils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Binding failed for port 8af4e5da-0239-4399-bc9e-79525f73ce64, please check neutron logs for more information. {{(pid=62183) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 843.491032] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.731s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.491134] env[62183]: DEBUG nova.objects.instance [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] [instance: 07899178-b53c-4b0e-877d-11774286212f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62183) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 843.495502] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Build of instance 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4 was re-scheduled: Binding failed for port 8af4e5da-0239-4399-bc9e-79525f73ce64, please check neutron logs for more information. {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 843.496111] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Unplugging VIFs for instance {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 843.496454] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.496739] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.497126] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.546752] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387028, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.559819] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5288fdce-8079-0023-d1fe-c50818852ef7, 'name': SearchDatastore_Task, 'duration_secs': 0.007887} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.560647] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d47e2e9f-87a0-4e55-8812-48e1117668be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.566513] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 843.566513] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]529990db-69f1-935a-3b4e-01b7536a8f6e" [ 843.566513] env[62183]: _type = "Task" [ 843.566513] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.573900] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529990db-69f1-935a-3b4e-01b7536a8f6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.717056] env[62183]: DEBUG nova.objects.instance [None req-f958ff6b-3a5d-4e67-9920-58b6ceae5a9f tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lazy-loading 'pci_requests' on Instance uuid 39683931-d0ff-4a5c-a4a2-792230ab0e3d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.721464] env[62183]: INFO nova.compute.manager [-] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Took 1.34 seconds to deallocate network for instance. [ 843.779413] env[62183]: INFO nova.compute.manager [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] [instance: 0dd90b88-ed92-4331-98c5-c7481ac1ae15] Took 1.02 seconds to deallocate network for instance. [ 843.898170] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.953568] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387029, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.019937] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.047394] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387028, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.559053} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.047685] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a/a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 844.047906] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.048182] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa601b67-31f0-4d06-86a6-3b38f5b32820 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.054859] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 844.054859] env[62183]: value = "task-1387030" [ 844.054859] env[62183]: _type = "Task" [ 844.054859] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.062666] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387030, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.078102] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529990db-69f1-935a-3b4e-01b7536a8f6e, 'name': SearchDatastore_Task, 'duration_secs': 0.024465} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.078434] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.078581] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93/4fa0beb2-3f36-48f4-996c-1da0bb891f93.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.078853] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c329cb3a-93dd-4d02-a887-159cd8044a12 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.085484] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 844.085484] env[62183]: value = "task-1387031" [ 844.085484] env[62183]: _type = "Task" [ 844.085484] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.094263] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387031, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.127787] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.220610] env[62183]: DEBUG nova.objects.base [None req-f958ff6b-3a5d-4e67-9920-58b6ceae5a9f tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Object Instance<39683931-d0ff-4a5c-a4a2-792230ab0e3d> lazy-loaded attributes: flavor,pci_requests {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 844.220610] env[62183]: DEBUG nova.network.neutron [None req-f958ff6b-3a5d-4e67-9920-58b6ceae5a9f tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 844.228078] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.303690] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f958ff6b-3a5d-4e67-9920-58b6ceae5a9f tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "interface-39683931-d0ff-4a5c-a4a2-792230ab0e3d-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.092s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.451962] env[62183]: DEBUG oslo_vmware.api [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387029, 'name': PowerOnVM_Task, 'duration_secs': 1.011065} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.452342] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 844.452547] env[62183]: INFO nova.compute.manager [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Took 8.22 seconds to spawn the instance on the hypervisor. [ 844.452725] env[62183]: DEBUG nova.compute.manager [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 844.453489] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6732ba-b21f-4f33-a81d-f3cb69fb40ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.507402] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7d8cadb-6c13-4211-a9de-ff5195973027 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.509262] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.080s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.510819] env[62183]: INFO nova.compute.claims [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.565562] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387030, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069663} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.565843] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 844.566673] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3b07d9-5117-4296-8017-754a51e5765b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.591076] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a/a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.591776] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6a90e02-ef5d-44b4-aeed-b6b151e31967 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.616036] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387031, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.617274] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 844.617274] env[62183]: value = "task-1387032" [ 844.617274] env[62183]: _type = "Task" [ 844.617274] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.625904] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387032, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.630672] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.630959] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62183) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 844.631259] env[62183]: DEBUG nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.631466] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.649242] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.809620] env[62183]: INFO nova.scheduler.client.report [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Deleted allocations for instance 0dd90b88-ed92-4331-98c5-c7481ac1ae15 [ 844.969590] env[62183]: INFO nova.compute.manager [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Took 28.40 seconds to build instance. [ 845.101582] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387031, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.849923} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.101809] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93/4fa0beb2-3f36-48f4-996c-1da0bb891f93.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 845.102081] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 845.102266] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa6f4064-809e-4ee0-9f82-7c586cca60b7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.108482] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 845.108482] env[62183]: value = "task-1387033" [ 845.108482] env[62183]: _type = "Task" [ 845.108482] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.116201] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387033, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.127779] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387032, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.152714] env[62183]: DEBUG nova.network.neutron [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.318212] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c3105a6b-cca4-4af5-a2e6-b94b2eecb22f tempest-ServersNegativeTestJSON-615521801 tempest-ServersNegativeTestJSON-615521801-project-member] Lock "0dd90b88-ed92-4331-98c5-c7481ac1ae15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.219s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.471505] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b365b45b-6d80-449b-b676-7af5f52efdb3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.702s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.617531] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387033, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065436} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.619967] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.621106] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5261b785-bbcb-44fd-a4a9-ccdbd5bcf5ac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.631010] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387032, 'name': ReconfigVM_Task, 'duration_secs': 0.916159} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.641101] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Reconfigured VM instance instance-00000045 to attach disk [datastore2] a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a/a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.650658] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93/4fa0beb2-3f36-48f4-996c-1da0bb891f93.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.651136] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-54e5a22d-9f5e-4945-90aa-02ba65d96b6e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.652985] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-325b0211-e26b-4058-9119-9a63923f8aff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.669409] env[62183]: INFO nova.compute.manager [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4] Took 1.04 seconds to deallocate network for instance. [ 845.677886] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 845.677886] env[62183]: value = "task-1387034" [ 845.677886] env[62183]: _type = "Task" [ 845.677886] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.679281] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 845.679281] env[62183]: value = "task-1387035" [ 845.679281] env[62183]: _type = "Task" [ 845.679281] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.693034] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387034, 'name': Rename_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.696019] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387035, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.753159] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2427d0-e558-46d6-89a9-18d5961566f3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.763111] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35697c53-91d1-4392-83e0-6fdea227499b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.794702] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b6da72-1168-4143-960f-1ac79cccbd6c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.802599] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de677cb-d51b-4786-b6f1-71084df2c333 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.816229] env[62183]: DEBUG nova.compute.provider_tree [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.824876] env[62183]: DEBUG nova.compute.manager [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 846.191598] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387035, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.194529] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387034, 'name': Rename_Task, 'duration_secs': 0.325178} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.194801] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 846.195017] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4c22d0d-055f-46c5-8f59-04d54f4b2322 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.200730] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 846.200730] env[62183]: value = "task-1387036" [ 846.200730] env[62183]: _type = "Task" [ 846.200730] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.208326] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387036, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.319262] env[62183]: DEBUG nova.scheduler.client.report [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 846.348922] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.451576] env[62183]: DEBUG oslo_concurrency.lockutils [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquiring lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.451576] env[62183]: DEBUG oslo_concurrency.lockutils [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.451576] env[62183]: DEBUG oslo_concurrency.lockutils [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquiring lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.451576] env[62183]: DEBUG oslo_concurrency.lockutils [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.451847] env[62183]: DEBUG oslo_concurrency.lockutils [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.454024] env[62183]: INFO nova.compute.manager [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Terminating instance [ 846.455809] env[62183]: DEBUG nova.compute.manager [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.456174] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.457028] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2934a58e-b936-4299-8117-d0cf3b7a5591 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.467019] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.467019] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cc797108-cb44-4b6e-97a8-550de4cf8497 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.472506] env[62183]: DEBUG oslo_vmware.api [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 846.472506] env[62183]: value = "task-1387037" [ 846.472506] env[62183]: _type = "Task" [ 846.472506] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.481694] env[62183]: DEBUG oslo_vmware.api [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.519144] env[62183]: DEBUG nova.compute.manager [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.520061] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f653dde-d54a-4ff1-8a31-8638a4824fc5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.694949] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387035, 'name': ReconfigVM_Task, 'duration_secs': 0.909432} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.694949] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93/4fa0beb2-3f36-48f4-996c-1da0bb891f93.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.695076] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-add0b330-9ea1-4c0d-8c0b-debf225d2282 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.701592] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 846.701592] env[62183]: value = "task-1387038" [ 846.701592] env[62183]: _type = "Task" [ 846.701592] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.705659] env[62183]: INFO nova.scheduler.client.report [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted allocations for instance 6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4 [ 846.719303] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387036, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.722603] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387038, 'name': Rename_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.825269] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.825844] env[62183]: DEBUG nova.compute.manager [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 846.829321] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.007s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.829610] env[62183]: DEBUG nova.objects.instance [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lazy-loading 'resources' on Instance uuid 07899178-b53c-4b0e-877d-11774286212f {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.984166] env[62183]: DEBUG oslo_vmware.api [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.032052] env[62183]: INFO nova.compute.manager [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] instance snapshotting [ 847.036247] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eefe824-85b7-4d15-9f6f-e68ee804585e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.059244] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4bf3ac-2bba-4ea8-b2a0-6397f8044242 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.214895] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387036, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.220148] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3eac4a9a-380c-4b04-a63a-450cc3da33ad tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "6aa3cfe4-42a4-4e5f-89cb-db4bbab791a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.699s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.220148] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387038, 'name': Rename_Task, 'duration_secs': 0.258989} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.220148] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.220148] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3be0f60c-f837-4571-b97c-e2b3a44e9e4c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.225539] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 847.225539] env[62183]: value = "task-1387039" [ 847.225539] env[62183]: _type = "Task" [ 847.225539] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.234429] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387039, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.333632] env[62183]: DEBUG nova.compute.utils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.338446] env[62183]: DEBUG nova.compute.manager [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 847.338725] env[62183]: DEBUG nova.network.neutron [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 847.400211] env[62183]: DEBUG nova.policy [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ee0f229ad7e4628bb17d089f15057fb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f32cc07823ca4f2aaba15ad9e2a5c761', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 847.489315] env[62183]: DEBUG oslo_vmware.api [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387037, 'name': PowerOffVM_Task, 'duration_secs': 0.694592} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.489428] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.489757] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.490202] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b1a8c7c-a3c9-431d-b15e-efe1686e2767 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.558549] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 847.559518] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 847.559518] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Deleting the datastore file [datastore1] 39683931-d0ff-4a5c-a4a2-792230ab0e3d {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.559518] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-961112f0-19d7-49a1-8761-e613e48f1ce0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.565135] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad2b9cd-0c31-4adf-bc05-12fdeb86792f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.569096] env[62183]: DEBUG oslo_vmware.api [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for the task: (returnval){ [ 847.569096] env[62183]: value = "task-1387041" [ 847.569096] env[62183]: _type = "Task" [ 847.569096] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.575536] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Creating Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 847.575843] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4482478d-2e11-4858-958e-a51958de441b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.578550] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd0d87a-ad5c-4dfc-9bf5-fe640a38fa57 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.584476] env[62183]: DEBUG oslo_vmware.api [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387041, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.620333] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4d885c-e383-40f3-ba35-ecebf848f7e8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.623151] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 847.623151] env[62183]: value = "task-1387042" [ 847.623151] env[62183]: _type = "Task" [ 847.623151] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.629783] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d86feb6-b3e6-4966-ab6c-dde88201acb8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.636310] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387042, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.646603] env[62183]: DEBUG nova.compute.provider_tree [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.713721] env[62183]: DEBUG oslo_vmware.api [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387036, 'name': PowerOnVM_Task, 'duration_secs': 1.227335} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.714012] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.714373] env[62183]: INFO nova.compute.manager [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Took 9.17 seconds to spawn the instance on the hypervisor. [ 847.715747] env[62183]: DEBUG nova.compute.manager [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.715747] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9477f846-c188-4b35-8dbd-1a85b2cc7bc6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.730105] env[62183]: DEBUG nova.network.neutron [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Successfully created port: d3884443-e294-4422-9ebe-924202ddaeab {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 847.740634] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387039, 'name': PowerOnVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.840677] env[62183]: DEBUG nova.compute.manager [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 848.083962] env[62183]: DEBUG oslo_vmware.api [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Task: {'id': task-1387041, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169115} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.084960] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 848.085197] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 848.085380] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 848.089018] env[62183]: INFO nova.compute.manager [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Took 1.63 seconds to destroy the instance on the hypervisor. [ 848.089018] env[62183]: DEBUG oslo.service.loopingcall [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.089018] env[62183]: DEBUG nova.compute.manager [-] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 848.089018] env[62183]: DEBUG nova.network.neutron [-] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 848.134205] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387042, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.150464] env[62183]: DEBUG nova.scheduler.client.report [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 848.239409] env[62183]: INFO nova.compute.manager [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Took 28.90 seconds to build instance. [ 848.244677] env[62183]: DEBUG oslo_vmware.api [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387039, 'name': PowerOnVM_Task, 'duration_secs': 0.607654} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.245744] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.245744] env[62183]: INFO nova.compute.manager [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Took 6.40 seconds to spawn the instance on the hypervisor. [ 848.245744] env[62183]: DEBUG nova.compute.manager [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.246679] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009741f1-2b54-4d33-a364-91f6c286befb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.635807] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387042, 'name': CreateSnapshot_Task, 'duration_secs': 0.875869} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.636596] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Created Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 848.637361] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fa0b7e-e3e1-46f8-96ed-23373e6b20d5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.649242] env[62183]: DEBUG nova.compute.manager [req-d87e8ae4-27ee-4197-b522-72d0a33758d4 req-f8cb46f7-e3d9-4e08-98f9-5a9fa4943cac service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Received event network-vif-deleted-eb7f6201-df05-4a22-b815-57d19d54b5fa {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.649415] env[62183]: INFO nova.compute.manager [req-d87e8ae4-27ee-4197-b522-72d0a33758d4 req-f8cb46f7-e3d9-4e08-98f9-5a9fa4943cac service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Neutron deleted interface eb7f6201-df05-4a22-b815-57d19d54b5fa; detaching it from the instance and deleting it from the info cache [ 848.649553] env[62183]: DEBUG nova.network.neutron [req-d87e8ae4-27ee-4197-b522-72d0a33758d4 req-f8cb46f7-e3d9-4e08-98f9-5a9fa4943cac service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.659849] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.830s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.663460] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.250s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.668120] env[62183]: INFO nova.compute.claims [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.689012] env[62183]: INFO nova.scheduler.client.report [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Deleted allocations for instance 07899178-b53c-4b0e-877d-11774286212f [ 848.745829] env[62183]: DEBUG oslo_concurrency.lockutils [None req-225cf32f-3811-4019-8d63-761b030ddc42 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.952s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.773084] env[62183]: INFO nova.compute.manager [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Took 27.65 seconds to build instance. [ 848.857054] env[62183]: DEBUG nova.compute.manager [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 848.892034] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 848.892160] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 848.892260] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 848.892456] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 848.892609] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 848.892739] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 848.892946] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 848.893192] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 848.893402] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 848.893601] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 848.893845] env[62183]: DEBUG nova.virt.hardware [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.894809] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b3acb7-aebf-42b7-afaf-9fb13ad468ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.903762] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c1b02d-2083-456f-9ac5-f8b2abffb47d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.919600] env[62183]: DEBUG nova.compute.manager [req-1f987238-f071-4a25-b5a8-ad5c56d4d51a req-8449eb57-bed5-4663-9f15-cf831f869562 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Received event network-changed-bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.919786] env[62183]: DEBUG nova.compute.manager [req-1f987238-f071-4a25-b5a8-ad5c56d4d51a req-8449eb57-bed5-4663-9f15-cf831f869562 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Refreshing instance network info cache due to event network-changed-bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 848.919997] env[62183]: DEBUG oslo_concurrency.lockutils [req-1f987238-f071-4a25-b5a8-ad5c56d4d51a req-8449eb57-bed5-4663-9f15-cf831f869562 service nova] Acquiring lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.920305] env[62183]: DEBUG oslo_concurrency.lockutils [req-1f987238-f071-4a25-b5a8-ad5c56d4d51a req-8449eb57-bed5-4663-9f15-cf831f869562 service nova] Acquired lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.920584] env[62183]: DEBUG nova.network.neutron [req-1f987238-f071-4a25-b5a8-ad5c56d4d51a req-8449eb57-bed5-4663-9f15-cf831f869562 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Refreshing network info cache for port bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.002852] env[62183]: DEBUG nova.network.neutron [-] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.159623] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Creating linked-clone VM from snapshot {{(pid=62183) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 849.160297] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d604ab0b-6537-415e-935e-ca2f7693375a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.164111] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f121119-8f27-46d2-8fba-ca8a8d1b35d1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.174181] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8a4818-ae68-48f5-800d-e975ba953b12 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.189283] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 849.189283] env[62183]: value = "task-1387043" [ 849.189283] env[62183]: _type = "Task" [ 849.189283] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.199601] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387043, 'name': CloneVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.200511] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8728e0a7-ef97-4c0c-9d9b-14116950fa69 tempest-ServerShowV254Test-384847156 tempest-ServerShowV254Test-384847156-project-member] Lock "07899178-b53c-4b0e-877d-11774286212f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.100s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.207951] env[62183]: DEBUG nova.compute.manager [req-d87e8ae4-27ee-4197-b522-72d0a33758d4 req-f8cb46f7-e3d9-4e08-98f9-5a9fa4943cac service nova] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Detach interface failed, port_id=eb7f6201-df05-4a22-b815-57d19d54b5fa, reason: Instance 39683931-d0ff-4a5c-a4a2-792230ab0e3d could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 849.259642] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "843d35ff-6871-4e2a-8e03-d7229a0d8246" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.260859] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.274956] env[62183]: DEBUG oslo_concurrency.lockutils [None req-526674d4-b0d0-4f35-9faa-624fae1b744a tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "4fa0beb2-3f36-48f4-996c-1da0bb891f93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.903s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.508234] env[62183]: INFO nova.compute.manager [-] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Took 1.42 seconds to deallocate network for instance. [ 849.700802] env[62183]: DEBUG nova.network.neutron [req-1f987238-f071-4a25-b5a8-ad5c56d4d51a req-8449eb57-bed5-4663-9f15-cf831f869562 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updated VIF entry in instance network info cache for port bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 849.701611] env[62183]: DEBUG nova.network.neutron [req-1f987238-f071-4a25-b5a8-ad5c56d4d51a req-8449eb57-bed5-4663-9f15-cf831f869562 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.710480] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387043, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.762027] env[62183]: DEBUG nova.compute.manager [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 849.833664] env[62183]: DEBUG nova.network.neutron [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Successfully updated port: d3884443-e294-4422-9ebe-924202ddaeab {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.865323] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8aae0a-d1f6-44a2-9e62-379124775a81 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.873243] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ac9bc6-5687-440b-881c-29d3f041377e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.903608] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a290f96a-ef2b-4a78-8690-acbc0fec4cb7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.911765] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd535fc-a77e-4cde-b045-c8ea0fb9fd37 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.925195] env[62183]: DEBUG nova.compute.provider_tree [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.975022] env[62183]: INFO nova.compute.manager [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Rebuilding instance [ 850.016228] env[62183]: DEBUG oslo_concurrency.lockutils [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.017709] env[62183]: DEBUG nova.compute.manager [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.017928] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300b63ab-ba9a-470f-8e41-0d05de6e9972 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.203645] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387043, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.205335] env[62183]: DEBUG oslo_concurrency.lockutils [req-1f987238-f071-4a25-b5a8-ad5c56d4d51a req-8449eb57-bed5-4663-9f15-cf831f869562 service nova] Releasing lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.284111] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.339720] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquiring lock "refresh_cache-e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.339889] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquired lock "refresh_cache-e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.339999] env[62183]: DEBUG nova.network.neutron [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.428396] env[62183]: DEBUG nova.scheduler.client.report [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 850.529360] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.529709] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97691ff6-ea86-455a-a7ea-5bd54fc33d48 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.537432] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 850.537432] env[62183]: value = "task-1387044" [ 850.537432] env[62183]: _type = "Task" [ 850.537432] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.548712] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387044, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.704338] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387043, 'name': CloneVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.872389] env[62183]: DEBUG nova.network.neutron [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.934154] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.934781] env[62183]: DEBUG nova.compute.manager [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 850.938247] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.033s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.940266] env[62183]: INFO nova.compute.claims [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.950509] env[62183]: DEBUG nova.compute.manager [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Received event network-vif-plugged-d3884443-e294-4422-9ebe-924202ddaeab {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.950722] env[62183]: DEBUG oslo_concurrency.lockutils [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] Acquiring lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.950934] env[62183]: DEBUG oslo_concurrency.lockutils [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] Lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.951114] env[62183]: DEBUG oslo_concurrency.lockutils [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] Lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.951313] env[62183]: DEBUG nova.compute.manager [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] No waiting events found dispatching network-vif-plugged-d3884443-e294-4422-9ebe-924202ddaeab {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.951446] env[62183]: WARNING nova.compute.manager [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Received unexpected event network-vif-plugged-d3884443-e294-4422-9ebe-924202ddaeab for instance with vm_state building and task_state spawning. [ 850.951604] env[62183]: DEBUG nova.compute.manager [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Received event network-changed-d3884443-e294-4422-9ebe-924202ddaeab {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.951780] env[62183]: DEBUG nova.compute.manager [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Refreshing instance network info cache due to event network-changed-d3884443-e294-4422-9ebe-924202ddaeab. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.951926] env[62183]: DEBUG oslo_concurrency.lockutils [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] Acquiring lock "refresh_cache-e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.023654] env[62183]: DEBUG nova.network.neutron [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Updating instance_info_cache with network_info: [{"id": "d3884443-e294-4422-9ebe-924202ddaeab", "address": "fa:16:3e:80:4b:90", "network": {"id": "afe869c4-c854-471d-81a8-7a6b8158f3fc", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1165865747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f32cc07823ca4f2aaba15ad9e2a5c761", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d62c1cf-f39a-4626-9552-f1e13c692636", "external-id": "nsx-vlan-transportzone-748", "segmentation_id": 748, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3884443-e2", "ovs_interfaceid": "d3884443-e294-4422-9ebe-924202ddaeab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.046913] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387044, 'name': PowerOffVM_Task, 'duration_secs': 0.210321} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.047227] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 851.047450] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 851.048201] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e85420-b5a8-49bc-9dcf-9323d75a21db {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.054907] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 851.055125] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e13cbcda-5939-4e38-8daf-a9dc916e0e93 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.077017] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.077256] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.077434] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Deleting the datastore file [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.077682] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eaa130cb-5117-4563-97af-e8e7f190c626 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.083259] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 851.083259] env[62183]: value = "task-1387046" [ 851.083259] env[62183]: _type = "Task" [ 851.083259] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.091097] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387046, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.204437] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387043, 'name': CloneVM_Task, 'duration_secs': 1.552239} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.204712] env[62183]: INFO nova.virt.vmwareapi.vmops [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Created linked-clone VM from snapshot [ 851.205433] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1f12ff-267f-46cf-baab-2ee45417b502 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.212716] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Uploading image 2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83 {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 851.232916] env[62183]: DEBUG oslo_vmware.rw_handles [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 851.232916] env[62183]: value = "vm-294443" [ 851.232916] env[62183]: _type = "VirtualMachine" [ 851.232916] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 851.233200] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-47b5f2e8-8ea3-45a6-b63c-79a1fe91dfdd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.239855] env[62183]: DEBUG oslo_vmware.rw_handles [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lease: (returnval){ [ 851.239855] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52403165-e3dd-a7c0-453e-8a8083d047db" [ 851.239855] env[62183]: _type = "HttpNfcLease" [ 851.239855] env[62183]: } obtained for exporting VM: (result){ [ 851.239855] env[62183]: value = "vm-294443" [ 851.239855] env[62183]: _type = "VirtualMachine" [ 851.239855] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 851.240095] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the lease: (returnval){ [ 851.240095] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52403165-e3dd-a7c0-453e-8a8083d047db" [ 851.240095] env[62183]: _type = "HttpNfcLease" [ 851.240095] env[62183]: } to be ready. {{(pid=62183) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 851.246307] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 851.246307] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52403165-e3dd-a7c0-453e-8a8083d047db" [ 851.246307] env[62183]: _type = "HttpNfcLease" [ 851.246307] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 851.445837] env[62183]: DEBUG nova.compute.utils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.447367] env[62183]: DEBUG nova.compute.manager [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.447818] env[62183]: DEBUG nova.network.neutron [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.517086] env[62183]: DEBUG nova.policy [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4b92513f568466e81075af3fa4604fa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88022032e6e04a4f96bc49c2ca5ede29', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.525490] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Releasing lock "refresh_cache-e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.525781] env[62183]: DEBUG nova.compute.manager [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Instance network_info: |[{"id": "d3884443-e294-4422-9ebe-924202ddaeab", "address": "fa:16:3e:80:4b:90", "network": {"id": "afe869c4-c854-471d-81a8-7a6b8158f3fc", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1165865747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f32cc07823ca4f2aaba15ad9e2a5c761", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d62c1cf-f39a-4626-9552-f1e13c692636", "external-id": "nsx-vlan-transportzone-748", "segmentation_id": 748, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3884443-e2", "ovs_interfaceid": "d3884443-e294-4422-9ebe-924202ddaeab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 851.526075] env[62183]: DEBUG oslo_concurrency.lockutils [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] Acquired lock "refresh_cache-e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.526280] env[62183]: DEBUG nova.network.neutron [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Refreshing network info cache for port d3884443-e294-4422-9ebe-924202ddaeab {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.527340] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:4b:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d62c1cf-f39a-4626-9552-f1e13c692636', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3884443-e294-4422-9ebe-924202ddaeab', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.534737] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Creating folder: Project (f32cc07823ca4f2aaba15ad9e2a5c761). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.537740] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c0b24287-6374-4e25-abfc-ee30b6cf90b1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.551027] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Created folder: Project (f32cc07823ca4f2aaba15ad9e2a5c761) in parent group-v294392. [ 851.551148] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Creating folder: Instances. Parent ref: group-v294444. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 851.551376] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-813d55ac-0d34-4026-a08c-7e53743809e8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.561448] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Created folder: Instances in parent group-v294444. [ 851.561689] env[62183]: DEBUG oslo.service.loopingcall [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.561925] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 851.562150] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22d8b5e0-2ad3-43b0-beb1-abec4e19a330 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.583136] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.583136] env[62183]: value = "task-1387050" [ 851.583136] env[62183]: _type = "Task" [ 851.583136] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.595451] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387050, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.598380] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387046, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10732} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.598614] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.598789] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.598963] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.749199] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 851.749199] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52403165-e3dd-a7c0-453e-8a8083d047db" [ 851.749199] env[62183]: _type = "HttpNfcLease" [ 851.749199] env[62183]: } is ready. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 851.749549] env[62183]: DEBUG oslo_vmware.rw_handles [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 851.749549] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52403165-e3dd-a7c0-453e-8a8083d047db" [ 851.749549] env[62183]: _type = "HttpNfcLease" [ 851.749549] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 851.750281] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf972d1-a5aa-4163-9424-95ad427c0667 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.759270] env[62183]: DEBUG oslo_vmware.rw_handles [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52614986-bf30-669f-0691-ea649b1c59e6/disk-0.vmdk from lease info. {{(pid=62183) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 851.759565] env[62183]: DEBUG oslo_vmware.rw_handles [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52614986-bf30-669f-0691-ea649b1c59e6/disk-0.vmdk for reading. {{(pid=62183) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 851.823295] env[62183]: DEBUG nova.network.neutron [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Updated VIF entry in instance network info cache for port d3884443-e294-4422-9ebe-924202ddaeab. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.823295] env[62183]: DEBUG nova.network.neutron [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Updating instance_info_cache with network_info: [{"id": "d3884443-e294-4422-9ebe-924202ddaeab", "address": "fa:16:3e:80:4b:90", "network": {"id": "afe869c4-c854-471d-81a8-7a6b8158f3fc", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1165865747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f32cc07823ca4f2aaba15ad9e2a5c761", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d62c1cf-f39a-4626-9552-f1e13c692636", "external-id": "nsx-vlan-transportzone-748", "segmentation_id": 748, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3884443-e2", "ovs_interfaceid": "d3884443-e294-4422-9ebe-924202ddaeab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.851856] env[62183]: DEBUG nova.network.neutron [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Successfully created port: 3d687443-41fd-44ff-8981-35319b79d88a {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.858011] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d2d55529-c45e-4a31-b3c7-39b46c2afb1d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.950838] env[62183]: DEBUG nova.compute.manager [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.098325] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387050, 'name': CreateVM_Task, 'duration_secs': 0.292624} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.098325] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 852.099517] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.099517] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.099517] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 852.099906] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db664033-2765-4bc9-9fc5-bb43574bc760 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.110301] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 852.110301] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]529d5197-058b-f84a-6785-e52894c18f22" [ 852.110301] env[62183]: _type = "Task" [ 852.110301] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.125421] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529d5197-058b-f84a-6785-e52894c18f22, 'name': SearchDatastore_Task, 'duration_secs': 0.010608} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.125421] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.125421] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.125421] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.125667] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.125667] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.128510] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55b1e48c-7035-42d8-8b3c-47a8b8fc8289 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.137310] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.137504] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.138364] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43b5ae6a-0793-4d8e-9467-165e2fdb2ae4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.147493] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 852.147493] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52dc41a4-7369-9180-3c7c-39c23f399e20" [ 852.147493] env[62183]: _type = "Task" [ 852.147493] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.156600] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52dc41a4-7369-9180-3c7c-39c23f399e20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.169963] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0d4663-a2e5-436c-b135-80a526b329f5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.177457] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a22bfe-dce9-4fb2-b109-8eecd7ea4e9a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.210614] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ce6fbc-b670-4017-8685-d2fc37986ac6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.219947] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2ab44f-05dd-4e25-85ea-724e04514524 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.235066] env[62183]: DEBUG nova.compute.provider_tree [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.326247] env[62183]: DEBUG oslo_concurrency.lockutils [req-97098ec0-4db1-403f-99db-b77c38fc9950 req-f5067463-5dfc-4243-aec2-a6989284145c service nova] Releasing lock "refresh_cache-e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.633722] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.634284] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.634563] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.634764] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.635060] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.635374] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.635726] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.636301] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.636592] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.636994] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.637117] env[62183]: DEBUG nova.virt.hardware [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.638553] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf99be14-7efc-48fa-901a-aaff8fc0c11d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.647612] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a654dd60-01cd-4674-bf9a-fe622868f2fb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.664982] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Instance VIF info [] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.670649] env[62183]: DEBUG oslo.service.loopingcall [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.674423] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.674751] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52dc41a4-7369-9180-3c7c-39c23f399e20, 'name': SearchDatastore_Task, 'duration_secs': 0.008761} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.674960] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58261fc1-8c83-4bb8-8720-e91c45255fb0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.687522] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-855d6960-1820-414a-9c65-d614e7dfbc94 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.694488] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 852.694488] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52110d16-aea2-e01a-1e53-eb28eb27bc9a" [ 852.694488] env[62183]: _type = "Task" [ 852.694488] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.695867] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.695867] env[62183]: value = "task-1387051" [ 852.695867] env[62183]: _type = "Task" [ 852.695867] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.707877] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52110d16-aea2-e01a-1e53-eb28eb27bc9a, 'name': SearchDatastore_Task, 'duration_secs': 0.010464} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.711356] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.711830] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] e3145bd7-85b2-4cc7-9d97-3e36a59b89cb/e3145bd7-85b2-4cc7-9d97-3e36a59b89cb.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.712635] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387051, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.712635] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0cea7e3-ef03-4d71-93d5-a186bfcc2c2e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.719615] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 852.719615] env[62183]: value = "task-1387052" [ 852.719615] env[62183]: _type = "Task" [ 852.719615] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.729865] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387052, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.738992] env[62183]: DEBUG nova.scheduler.client.report [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 852.964209] env[62183]: DEBUG nova.compute.manager [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 852.991584] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.991857] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.992280] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.992558] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.992720] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.992875] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.993314] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.993670] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.993962] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.994194] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.994393] env[62183]: DEBUG nova.virt.hardware [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.996569] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a2dbee-c64b-4af7-b118-614c4a857178 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.007167] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70695b93-0849-4853-a2de-f880b62b2515 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.210353] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387051, 'name': CreateVM_Task, 'duration_secs': 0.513048} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.210528] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 853.212116] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.212116] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.212116] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 853.212116] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c351bef5-d7ac-4d2a-9a1b-ee709c298ddf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.217500] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 853.217500] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]522d3459-e3a7-b908-2427-9ff2ed2e2d5f" [ 853.217500] env[62183]: _type = "Task" [ 853.217500] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.230648] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]522d3459-e3a7-b908-2427-9ff2ed2e2d5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.235183] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387052, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.248486] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.249455] env[62183]: DEBUG nova.compute.manager [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 853.252274] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.524s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.254123] env[62183]: INFO nova.compute.claims [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.312502] env[62183]: DEBUG nova.compute.manager [req-3cc87b93-1d08-4a63-8f9b-b262436343ec req-ded96ce5-effc-472b-a6d9-003553b3969d service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Received event network-vif-plugged-3d687443-41fd-44ff-8981-35319b79d88a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.312808] env[62183]: DEBUG oslo_concurrency.lockutils [req-3cc87b93-1d08-4a63-8f9b-b262436343ec req-ded96ce5-effc-472b-a6d9-003553b3969d service nova] Acquiring lock "a948464b-63aa-4bc8-9885-228049e96d37-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.313131] env[62183]: DEBUG oslo_concurrency.lockutils [req-3cc87b93-1d08-4a63-8f9b-b262436343ec req-ded96ce5-effc-472b-a6d9-003553b3969d service nova] Lock "a948464b-63aa-4bc8-9885-228049e96d37-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.313356] env[62183]: DEBUG oslo_concurrency.lockutils [req-3cc87b93-1d08-4a63-8f9b-b262436343ec req-ded96ce5-effc-472b-a6d9-003553b3969d service nova] Lock "a948464b-63aa-4bc8-9885-228049e96d37-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.313580] env[62183]: DEBUG nova.compute.manager [req-3cc87b93-1d08-4a63-8f9b-b262436343ec req-ded96ce5-effc-472b-a6d9-003553b3969d service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] No waiting events found dispatching network-vif-plugged-3d687443-41fd-44ff-8981-35319b79d88a {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.313889] env[62183]: WARNING nova.compute.manager [req-3cc87b93-1d08-4a63-8f9b-b262436343ec req-ded96ce5-effc-472b-a6d9-003553b3969d service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Received unexpected event network-vif-plugged-3d687443-41fd-44ff-8981-35319b79d88a for instance with vm_state building and task_state spawning. [ 853.412393] env[62183]: DEBUG nova.network.neutron [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Successfully updated port: 3d687443-41fd-44ff-8981-35319b79d88a {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.728836] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]522d3459-e3a7-b908-2427-9ff2ed2e2d5f, 'name': SearchDatastore_Task, 'duration_secs': 0.047918} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.729555] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.729800] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 853.730109] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.730283] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.730472] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 853.730746] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cca6c117-6117-4fd1-b67a-bc4ba3c8b94f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.735800] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387052, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.559461} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.736407] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] e3145bd7-85b2-4cc7-9d97-3e36a59b89cb/e3145bd7-85b2-4cc7-9d97-3e36a59b89cb.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 853.736620] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.736893] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8edef1f-c68b-46ac-ac18-e54765829d33 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.742318] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 853.742495] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 853.744170] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b76ba18c-e97f-46c1-a282-a34880e4666c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.746832] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 853.746832] env[62183]: value = "task-1387053" [ 853.746832] env[62183]: _type = "Task" [ 853.746832] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.751652] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 853.751652] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5273af0b-baf3-f4e5-e74a-b57e808652b6" [ 853.751652] env[62183]: _type = "Task" [ 853.751652] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.758612] env[62183]: DEBUG nova.compute.utils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.759903] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387053, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.762434] env[62183]: DEBUG nova.compute.manager [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.762615] env[62183]: DEBUG nova.network.neutron [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 853.769549] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5273af0b-baf3-f4e5-e74a-b57e808652b6, 'name': SearchDatastore_Task, 'duration_secs': 0.010722} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.771147] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2b8414c-4e69-40b6-9ed4-d3bd337fcbc6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.776835] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 853.776835] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5237964e-7928-f00d-2466-1b9458560a75" [ 853.776835] env[62183]: _type = "Task" [ 853.776835] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.785924] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5237964e-7928-f00d-2466-1b9458560a75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.819418] env[62183]: DEBUG nova.policy [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89ad04efd23b40e6a3eea5b3e6fd021e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b61c6a99cbda435481bb72f20929f03f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.914399] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.914560] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.914701] env[62183]: DEBUG nova.network.neutron [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 854.075545] env[62183]: DEBUG nova.network.neutron [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Successfully created port: d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.256490] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387053, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066667} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.256776] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.257890] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f3aea0-b4e3-405d-8d7b-6d42f1d89cc2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.275527] env[62183]: DEBUG nova.compute.manager [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 854.286518] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] e3145bd7-85b2-4cc7-9d97-3e36a59b89cb/e3145bd7-85b2-4cc7-9d97-3e36a59b89cb.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.287379] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7d8d6d6-fff9-49df-a5fd-d90607b2189d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.314157] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 854.314157] env[62183]: value = "task-1387054" [ 854.314157] env[62183]: _type = "Task" [ 854.314157] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.315113] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5237964e-7928-f00d-2466-1b9458560a75, 'name': SearchDatastore_Task, 'duration_secs': 0.008901} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.315113] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.315113] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93/4fa0beb2-3f36-48f4-996c-1da0bb891f93.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 854.318158] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-087e41cd-8778-43e8-9265-84e32ae529b5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.326349] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387054, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.330469] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 854.330469] env[62183]: value = "task-1387055" [ 854.330469] env[62183]: _type = "Task" [ 854.330469] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.340268] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387055, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.452520] env[62183]: DEBUG nova.network.neutron [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.489299] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f625d669-ffb1-42a3-9953-6146c84f541c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.497530] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1271e2-7c43-4888-8c17-a7e429522ab2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.535341] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2572b50c-f627-48a8-a9c7-c95ca3e80543 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.545524] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2987261d-a0c9-4e93-80b3-47389e73da4c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.560631] env[62183]: DEBUG nova.compute.provider_tree [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.649662] env[62183]: DEBUG nova.network.neutron [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [{"id": "3d687443-41fd-44ff-8981-35319b79d88a", "address": "fa:16:3e:91:31:9b", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d687443-41", "ovs_interfaceid": "3d687443-41fd-44ff-8981-35319b79d88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.826692] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387054, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.840658] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387055, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.063951] env[62183]: DEBUG nova.scheduler.client.report [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 855.152151] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.152475] env[62183]: DEBUG nova.compute.manager [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Instance network_info: |[{"id": "3d687443-41fd-44ff-8981-35319b79d88a", "address": "fa:16:3e:91:31:9b", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d687443-41", "ovs_interfaceid": "3d687443-41fd-44ff-8981-35319b79d88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.152926] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:31:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ff90ec9-3c7e-4e76-b409-fcf37fc588d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d687443-41fd-44ff-8981-35319b79d88a', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.161291] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating folder: Project (88022032e6e04a4f96bc49c2ca5ede29). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.161618] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65555047-abbd-494c-b43d-f9eb23281ad8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.175096] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Created folder: Project (88022032e6e04a4f96bc49c2ca5ede29) in parent group-v294392. [ 855.175096] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating folder: Instances. Parent ref: group-v294448. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.175096] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bfd2f4a2-3c39-4b6f-b6cc-f933171faa3c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.186513] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Created folder: Instances in parent group-v294448. [ 855.186873] env[62183]: DEBUG oslo.service.loopingcall [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.187122] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.187359] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbdbb9cb-9d0b-42f2-8b80-43c646a9b868 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.207883] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.207883] env[62183]: value = "task-1387058" [ 855.207883] env[62183]: _type = "Task" [ 855.207883] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.217857] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387058, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.293572] env[62183]: DEBUG nova.compute.manager [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 855.316190] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.316509] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.316692] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.316898] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.317076] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.317245] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.317474] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.317644] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.317824] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.317995] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.318242] env[62183]: DEBUG nova.virt.hardware [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.319271] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093e8f0e-996a-4aef-ba25-891b4445ba5b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.333430] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608e841f-9120-450b-a7c2-e25cf719d654 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.338175] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387054, 'name': ReconfigVM_Task, 'duration_secs': 0.698471} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.344115] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Reconfigured VM instance instance-00000047 to attach disk [datastore2] e3145bd7-85b2-4cc7-9d97-3e36a59b89cb/e3145bd7-85b2-4cc7-9d97-3e36a59b89cb.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.345651] env[62183]: DEBUG nova.compute.manager [req-a96cf3b8-c497-4d7a-8c01-11181524d749 req-9ac1418f-5cba-426a-9fe3-bfdc8d058cae service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Received event network-changed-3d687443-41fd-44ff-8981-35319b79d88a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.345867] env[62183]: DEBUG nova.compute.manager [req-a96cf3b8-c497-4d7a-8c01-11181524d749 req-9ac1418f-5cba-426a-9fe3-bfdc8d058cae service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Refreshing instance network info cache due to event network-changed-3d687443-41fd-44ff-8981-35319b79d88a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 855.346107] env[62183]: DEBUG oslo_concurrency.lockutils [req-a96cf3b8-c497-4d7a-8c01-11181524d749 req-9ac1418f-5cba-426a-9fe3-bfdc8d058cae service nova] Acquiring lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.346270] env[62183]: DEBUG oslo_concurrency.lockutils [req-a96cf3b8-c497-4d7a-8c01-11181524d749 req-9ac1418f-5cba-426a-9fe3-bfdc8d058cae service nova] Acquired lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.346450] env[62183]: DEBUG nova.network.neutron [req-a96cf3b8-c497-4d7a-8c01-11181524d749 req-9ac1418f-5cba-426a-9fe3-bfdc8d058cae service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Refreshing network info cache for port 3d687443-41fd-44ff-8981-35319b79d88a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.349453] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b479719-1125-4a8b-a88a-e455bc616476 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.366088] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387055, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.617377} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.367516] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93/4fa0beb2-3f36-48f4-996c-1da0bb891f93.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 855.367782] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 855.368174] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 855.368174] env[62183]: value = "task-1387059" [ 855.368174] env[62183]: _type = "Task" [ 855.368174] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.368701] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cfed87da-9c4d-4f63-a33d-4b6689f5b004 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.381411] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387059, 'name': Rename_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.382537] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 855.382537] env[62183]: value = "task-1387060" [ 855.382537] env[62183]: _type = "Task" [ 855.382537] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.393815] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387060, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.569971] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.318s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.570558] env[62183]: DEBUG nova.compute.manager [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 855.573515] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.675s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.575065] env[62183]: INFO nova.compute.claims [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.597482] env[62183]: DEBUG nova.network.neutron [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Successfully updated port: d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.718145] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387058, 'name': CreateVM_Task, 'duration_secs': 0.37085} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.718384] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.719099] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.719304] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.719643] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.719940] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62b0c693-2d07-4e22-9c91-cfc05f1c8bda {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.724616] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 855.724616] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]528eb16f-8bf2-f41f-c2d5-1e4229bddc94" [ 855.724616] env[62183]: _type = "Task" [ 855.724616] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.732622] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528eb16f-8bf2-f41f-c2d5-1e4229bddc94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.880179] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387059, 'name': Rename_Task, 'duration_secs': 0.137819} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.880545] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.880740] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec5eb749-2555-4cca-a3b2-cd343bdda207 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.887344] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 855.887344] env[62183]: value = "task-1387061" [ 855.887344] env[62183]: _type = "Task" [ 855.887344] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.893555] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387060, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088047} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.896241] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 855.897103] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c38d8f-8a46-45be-9e78-76fea813eede {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.902966] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387061, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.920468] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93/4fa0beb2-3f36-48f4-996c-1da0bb891f93.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.920840] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55862f8b-2806-4e23-8595-8a1f2cc8bae7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.940692] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 855.940692] env[62183]: value = "task-1387062" [ 855.940692] env[62183]: _type = "Task" [ 855.940692] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.950495] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387062, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.079786] env[62183]: DEBUG nova.compute.utils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.084115] env[62183]: DEBUG nova.compute.manager [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 856.084190] env[62183]: DEBUG nova.network.neutron [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 856.100027] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.100261] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.100439] env[62183]: DEBUG nova.network.neutron [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.152070] env[62183]: DEBUG nova.policy [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2e53beb806094cd98c694aaffdad3b99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e74d26cb8f5a47a48dc025b8058c5768', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 856.243203] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528eb16f-8bf2-f41f-c2d5-1e4229bddc94, 'name': SearchDatastore_Task, 'duration_secs': 0.008582} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.247670] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.248292] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.248624] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.248825] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.249089] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.249834] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da31e720-d0d3-4b5b-8c49-37547e339135 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.261232] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.261485] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 856.262551] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8dd3a3b8-5734-49bf-adc5-9e528b21bd1f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.271015] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 856.271015] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5255c4a5-c6cb-fa6c-f705-c48cab2d93bf" [ 856.271015] env[62183]: _type = "Task" [ 856.271015] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.284713] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5255c4a5-c6cb-fa6c-f705-c48cab2d93bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.340191] env[62183]: DEBUG nova.network.neutron [req-a96cf3b8-c497-4d7a-8c01-11181524d749 req-9ac1418f-5cba-426a-9fe3-bfdc8d058cae service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updated VIF entry in instance network info cache for port 3d687443-41fd-44ff-8981-35319b79d88a. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 856.340191] env[62183]: DEBUG nova.network.neutron [req-a96cf3b8-c497-4d7a-8c01-11181524d749 req-9ac1418f-5cba-426a-9fe3-bfdc8d058cae service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [{"id": "3d687443-41fd-44ff-8981-35319b79d88a", "address": "fa:16:3e:91:31:9b", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d687443-41", "ovs_interfaceid": "3d687443-41fd-44ff-8981-35319b79d88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.398159] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387061, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.450682] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387062, 'name': ReconfigVM_Task, 'duration_secs': 0.362494} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.451040] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93/4fa0beb2-3f36-48f4-996c-1da0bb891f93.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.451869] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d75031ab-ffb9-4c21-a028-f5e51e684887 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.458784] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 856.458784] env[62183]: value = "task-1387063" [ 856.458784] env[62183]: _type = "Task" [ 856.458784] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.467988] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387063, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.535191] env[62183]: DEBUG nova.network.neutron [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Successfully created port: 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.585293] env[62183]: DEBUG nova.compute.manager [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.639654] env[62183]: DEBUG nova.network.neutron [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.781570] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5255c4a5-c6cb-fa6c-f705-c48cab2d93bf, 'name': SearchDatastore_Task, 'duration_secs': 0.024833} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.782618] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79bb8d91-b1c8-4c55-8e47-d91092fd57c5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.790432] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 856.790432] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]527d687c-db4d-6b2b-085c-c26b07fe8511" [ 856.790432] env[62183]: _type = "Task" [ 856.790432] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.800062] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527d687c-db4d-6b2b-085c-c26b07fe8511, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.822079] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36abca9-929c-4e1d-9a56-126f8f4bb440 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.829197] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933ff7ad-f16e-4fe6-9dbc-031d7f5083f8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.862260] env[62183]: DEBUG oslo_concurrency.lockutils [req-a96cf3b8-c497-4d7a-8c01-11181524d749 req-9ac1418f-5cba-426a-9fe3-bfdc8d058cae service nova] Releasing lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.863357] env[62183]: DEBUG nova.network.neutron [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance_info_cache with network_info: [{"id": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "address": "fa:16:3e:2f:6f:39", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2f9a654-1f", "ovs_interfaceid": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.865039] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7af5de6-e7e8-4825-8cb9-704135e89bc7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.872492] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8870ae7-937d-4048-8098-3f0d60d2e109 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.888080] env[62183]: DEBUG nova.compute.provider_tree [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.897357] env[62183]: DEBUG oslo_vmware.api [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387061, 'name': PowerOnVM_Task, 'duration_secs': 0.511731} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.898151] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.898361] env[62183]: INFO nova.compute.manager [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Took 8.04 seconds to spawn the instance on the hypervisor. [ 856.898546] env[62183]: DEBUG nova.compute.manager [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.899322] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d36e6e-4e77-4517-ad21-e9f20322a4c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.969206] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387063, 'name': Rename_Task, 'duration_secs': 0.277576} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.969569] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.969816] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cba5879c-deef-4f42-9900-761c5094b0af {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.977054] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 856.977054] env[62183]: value = "task-1387064" [ 856.977054] env[62183]: _type = "Task" [ 856.977054] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.988197] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387064, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.302080] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527d687c-db4d-6b2b-085c-c26b07fe8511, 'name': SearchDatastore_Task, 'duration_secs': 0.028199} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.302399] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.302667] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] a948464b-63aa-4bc8-9885-228049e96d37/a948464b-63aa-4bc8-9885-228049e96d37.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 857.302942] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f241a4e4-ce31-4cc2-a575-3a198ce30a96 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.310881] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 857.310881] env[62183]: value = "task-1387065" [ 857.310881] env[62183]: _type = "Task" [ 857.310881] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.318795] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387065, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.369505] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.369952] env[62183]: DEBUG nova.compute.manager [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Instance network_info: |[{"id": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "address": "fa:16:3e:2f:6f:39", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2f9a654-1f", "ovs_interfaceid": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.371774] env[62183]: DEBUG nova.compute.manager [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Received event network-vif-plugged-d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.372078] env[62183]: DEBUG oslo_concurrency.lockutils [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] Acquiring lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.372387] env[62183]: DEBUG oslo_concurrency.lockutils [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.372630] env[62183]: DEBUG oslo_concurrency.lockutils [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.372937] env[62183]: DEBUG nova.compute.manager [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] No waiting events found dispatching network-vif-plugged-d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 857.373171] env[62183]: WARNING nova.compute.manager [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Received unexpected event network-vif-plugged-d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29 for instance with vm_state building and task_state spawning. [ 857.373417] env[62183]: DEBUG nova.compute.manager [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Received event network-changed-d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.373647] env[62183]: DEBUG nova.compute.manager [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Refreshing instance network info cache due to event network-changed-d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.373897] env[62183]: DEBUG oslo_concurrency.lockutils [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] Acquiring lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.374149] env[62183]: DEBUG oslo_concurrency.lockutils [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] Acquired lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.374399] env[62183]: DEBUG nova.network.neutron [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Refreshing network info cache for port d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 857.376968] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:6f:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2f9a654-1fa3-48a5-a9ba-c6de69dc5f29', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.386363] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Creating folder: Project (b61c6a99cbda435481bb72f20929f03f). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.388289] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06ba732d-3fbb-4938-b80f-c881e3fc29e3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.394251] env[62183]: DEBUG nova.scheduler.client.report [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 857.400470] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Created folder: Project (b61c6a99cbda435481bb72f20929f03f) in parent group-v294392. [ 857.400814] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Creating folder: Instances. Parent ref: group-v294451. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.401040] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d45e82e-e8c5-41f0-983a-4ac48ee526ce {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.419780] env[62183]: INFO nova.compute.manager [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Took 30.01 seconds to build instance. [ 857.421557] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Created folder: Instances in parent group-v294451. [ 857.421900] env[62183]: DEBUG oslo.service.loopingcall [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.422276] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 857.422520] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-40dece8f-ef08-4602-ad04-a991e4dbf231 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.444551] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.444551] env[62183]: value = "task-1387068" [ 857.444551] env[62183]: _type = "Task" [ 857.444551] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.453126] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387068, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.488246] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387064, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.597807] env[62183]: DEBUG nova.compute.manager [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 857.620753] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 857.621102] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 857.621205] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.621384] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 857.621539] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.621942] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 857.622331] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 857.622552] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 857.622755] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 857.622941] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 857.623171] env[62183]: DEBUG nova.virt.hardware [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.624461] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1d504a-f99c-455a-8384-cc367e65c030 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.633099] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7a37f2-5642-4561-9bce-31dc9ac4be14 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.820563] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387065, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.899565] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.900131] env[62183]: DEBUG nova.compute.manager [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.902924] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.675s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.903378] env[62183]: DEBUG nova.objects.instance [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lazy-loading 'resources' on Instance uuid 68e791b4-61db-4b6b-a30a-ccf5d6657643 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.923189] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ce1f3e91-96f6-4764-a978-fd42b03f806c tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.291s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.957517] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387068, 'name': CreateVM_Task, 'duration_secs': 0.427998} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.957593] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 857.958226] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.958401] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.958722] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.959226] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-089f6b0e-18c0-4dbf-95a0-4046bc6af878 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.963610] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 857.963610] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]524b751e-54c6-1420-4af0-d743e098309c" [ 857.963610] env[62183]: _type = "Task" [ 857.963610] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.973789] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]524b751e-54c6-1420-4af0-d743e098309c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.986468] env[62183]: DEBUG oslo_vmware.api [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387064, 'name': PowerOnVM_Task, 'duration_secs': 0.58971} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.987075] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.987296] env[62183]: DEBUG nova.compute.manager [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.988286] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502cad63-051d-47cb-9857-e06d7065e77b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.118053] env[62183]: DEBUG nova.network.neutron [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updated VIF entry in instance network info cache for port d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 858.118419] env[62183]: DEBUG nova.network.neutron [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance_info_cache with network_info: [{"id": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "address": "fa:16:3e:2f:6f:39", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2f9a654-1f", "ovs_interfaceid": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.327338] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387065, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.384847] env[62183]: DEBUG nova.compute.manager [req-f049ec7c-4274-4fae-9541-deab045dc220 req-9419670e-3d83-4360-9dd7-c475208fca2d service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received event network-vif-plugged-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.385149] env[62183]: DEBUG oslo_concurrency.lockutils [req-f049ec7c-4274-4fae-9541-deab045dc220 req-9419670e-3d83-4360-9dd7-c475208fca2d service nova] Acquiring lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.385382] env[62183]: DEBUG oslo_concurrency.lockutils [req-f049ec7c-4274-4fae-9541-deab045dc220 req-9419670e-3d83-4360-9dd7-c475208fca2d service nova] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.385879] env[62183]: DEBUG oslo_concurrency.lockutils [req-f049ec7c-4274-4fae-9541-deab045dc220 req-9419670e-3d83-4360-9dd7-c475208fca2d service nova] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.385879] env[62183]: DEBUG nova.compute.manager [req-f049ec7c-4274-4fae-9541-deab045dc220 req-9419670e-3d83-4360-9dd7-c475208fca2d service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] No waiting events found dispatching network-vif-plugged-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 858.385879] env[62183]: WARNING nova.compute.manager [req-f049ec7c-4274-4fae-9541-deab045dc220 req-9419670e-3d83-4360-9dd7-c475208fca2d service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received unexpected event network-vif-plugged-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 for instance with vm_state building and task_state spawning. [ 858.409198] env[62183]: DEBUG nova.compute.utils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.410484] env[62183]: DEBUG nova.compute.manager [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 858.410671] env[62183]: DEBUG nova.network.neutron [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 858.464216] env[62183]: DEBUG nova.policy [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be747fe0622548afa03792bb0a46c726', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c9dfd905ec14f2bbbe2b98692222f17', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.481095] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]524b751e-54c6-1420-4af0-d743e098309c, 'name': SearchDatastore_Task, 'duration_secs': 0.01456} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.481095] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.481095] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.481095] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.482732] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.482732] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.482732] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f89663df-0848-4c97-bb10-1dbb9dfda125 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.487570] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.487762] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.490948] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c922e18-fe8b-4756-8499-4eb96b8ae086 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.498181] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 858.498181] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]528d6e27-c35f-0071-75ab-1c0a706f2bf3" [ 858.498181] env[62183]: _type = "Task" [ 858.498181] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.517510] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528d6e27-c35f-0071-75ab-1c0a706f2bf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.518121] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.620627] env[62183]: DEBUG oslo_concurrency.lockutils [req-8eeb34cb-67b5-458f-85f5-9209c669ab6c req-f247982d-8dd9-4474-bdb6-83d0307bd92b service nova] Releasing lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.643054] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb2a3aa-eb9e-4401-b5e5-8ba1520a386c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.651629] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74651a3c-ecd2-4f8c-902d-f909f356bda1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.687856] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "4fa0beb2-3f36-48f4-996c-1da0bb891f93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.688212] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "4fa0beb2-3f36-48f4-996c-1da0bb891f93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.688471] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "4fa0beb2-3f36-48f4-996c-1da0bb891f93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.688697] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "4fa0beb2-3f36-48f4-996c-1da0bb891f93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.688876] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "4fa0beb2-3f36-48f4-996c-1da0bb891f93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.692205] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8fe8c6-d623-46b7-b744-11b0716543a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.696072] env[62183]: DEBUG nova.compute.manager [req-a8f75545-1706-49c3-8774-fd40b324b25b req-9f5fdf8f-3804-4f17-9a52-3328bd94b4fb service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Received event network-changed-d3884443-e294-4422-9ebe-924202ddaeab {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.696315] env[62183]: DEBUG nova.compute.manager [req-a8f75545-1706-49c3-8774-fd40b324b25b req-9f5fdf8f-3804-4f17-9a52-3328bd94b4fb service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Refreshing instance network info cache due to event network-changed-d3884443-e294-4422-9ebe-924202ddaeab. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.696650] env[62183]: DEBUG oslo_concurrency.lockutils [req-a8f75545-1706-49c3-8774-fd40b324b25b req-9f5fdf8f-3804-4f17-9a52-3328bd94b4fb service nova] Acquiring lock "refresh_cache-e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.696844] env[62183]: DEBUG oslo_concurrency.lockutils [req-a8f75545-1706-49c3-8774-fd40b324b25b req-9f5fdf8f-3804-4f17-9a52-3328bd94b4fb service nova] Acquired lock "refresh_cache-e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.697055] env[62183]: DEBUG nova.network.neutron [req-a8f75545-1706-49c3-8774-fd40b324b25b req-9f5fdf8f-3804-4f17-9a52-3328bd94b4fb service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Refreshing network info cache for port d3884443-e294-4422-9ebe-924202ddaeab {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.703019] env[62183]: INFO nova.compute.manager [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Terminating instance [ 858.703019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "refresh_cache-4fa0beb2-3f36-48f4-996c-1da0bb891f93" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.703019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquired lock "refresh_cache-4fa0beb2-3f36-48f4-996c-1da0bb891f93" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.703019] env[62183]: DEBUG nova.network.neutron [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.710479] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d64af42-438c-450a-9abc-61a3b0da9e12 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.726400] env[62183]: DEBUG nova.compute.provider_tree [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.798975] env[62183]: DEBUG nova.network.neutron [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Successfully created port: 4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.822947] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387065, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.913944] env[62183]: DEBUG nova.compute.manager [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.967255] env[62183]: DEBUG nova.network.neutron [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Successfully updated port: 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.009608] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528d6e27-c35f-0071-75ab-1c0a706f2bf3, 'name': SearchDatastore_Task, 'duration_secs': 0.01513} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.010528] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3eb0ddd-b689-48b0-8a1a-79c826876af8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.016205] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 859.016205] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52b98d81-2ea8-51a7-c99a-35988b8edcc3" [ 859.016205] env[62183]: _type = "Task" [ 859.016205] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.024696] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b98d81-2ea8-51a7-c99a-35988b8edcc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.221868] env[62183]: DEBUG nova.network.neutron [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 859.229702] env[62183]: DEBUG nova.scheduler.client.report [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 859.298659] env[62183]: DEBUG nova.network.neutron [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.325083] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387065, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.474352] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.474794] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.474794] env[62183]: DEBUG nova.network.neutron [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.490343] env[62183]: DEBUG nova.network.neutron [req-a8f75545-1706-49c3-8774-fd40b324b25b req-9f5fdf8f-3804-4f17-9a52-3328bd94b4fb service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Updated VIF entry in instance network info cache for port d3884443-e294-4422-9ebe-924202ddaeab. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.490740] env[62183]: DEBUG nova.network.neutron [req-a8f75545-1706-49c3-8774-fd40b324b25b req-9f5fdf8f-3804-4f17-9a52-3328bd94b4fb service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Updating instance_info_cache with network_info: [{"id": "d3884443-e294-4422-9ebe-924202ddaeab", "address": "fa:16:3e:80:4b:90", "network": {"id": "afe869c4-c854-471d-81a8-7a6b8158f3fc", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1165865747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f32cc07823ca4f2aaba15ad9e2a5c761", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d62c1cf-f39a-4626-9552-f1e13c692636", "external-id": "nsx-vlan-transportzone-748", "segmentation_id": 748, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3884443-e2", "ovs_interfaceid": "d3884443-e294-4422-9ebe-924202ddaeab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.528628] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b98d81-2ea8-51a7-c99a-35988b8edcc3, 'name': SearchDatastore_Task, 'duration_secs': 0.013586} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.528796] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.529131] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 503786ca-dba4-43c1-9a25-9f1cbac9a6a4/503786ca-dba4-43c1-9a25-9f1cbac9a6a4.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.529462] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f464d1b0-3050-4eec-988a-cf62ddf3b5bb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.537484] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 859.537484] env[62183]: value = "task-1387069" [ 859.537484] env[62183]: _type = "Task" [ 859.537484] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.548083] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387069, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.734355] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.737652] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.388s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.741018] env[62183]: INFO nova.compute.claims [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.762147] env[62183]: INFO nova.scheduler.client.report [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Deleted allocations for instance 68e791b4-61db-4b6b-a30a-ccf5d6657643 [ 859.801725] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Releasing lock "refresh_cache-4fa0beb2-3f36-48f4-996c-1da0bb891f93" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.802225] env[62183]: DEBUG nova.compute.manager [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 859.802510] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.803450] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7706a48c-7b1e-4eb1-96ef-b7de4322053c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.812942] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.812942] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71b70c63-b20b-41a6-98b6-3d8c16587467 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.824092] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387065, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.367489} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.825455] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] a948464b-63aa-4bc8-9885-228049e96d37/a948464b-63aa-4bc8-9885-228049e96d37.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.825698] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.826048] env[62183]: DEBUG oslo_vmware.api [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 859.826048] env[62183]: value = "task-1387070" [ 859.826048] env[62183]: _type = "Task" [ 859.826048] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.826265] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6a71311-b31d-4235-9ec7-1b3104103b20 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.837412] env[62183]: DEBUG oslo_vmware.api [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387070, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.838839] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 859.838839] env[62183]: value = "task-1387071" [ 859.838839] env[62183]: _type = "Task" [ 859.838839] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.847698] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387071, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.925212] env[62183]: DEBUG nova.compute.manager [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.950146] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.950504] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.950689] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.950892] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.951069] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.951237] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.951465] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.951907] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.951907] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.952046] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.952347] env[62183]: DEBUG nova.virt.hardware [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.953264] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef3e290-9550-4e2e-adc7-d0a2e78bf599 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.961746] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8f0934-02e6-4104-8564-024709c9db14 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.993129] env[62183]: DEBUG oslo_concurrency.lockutils [req-a8f75545-1706-49c3-8774-fd40b324b25b req-9f5fdf8f-3804-4f17-9a52-3328bd94b4fb service nova] Releasing lock "refresh_cache-e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.035232] env[62183]: DEBUG nova.network.neutron [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.051336] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387069, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.275643] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8b2ebd26-c1e6-4f58-b0fc-9292f1c5b236 tempest-InstanceActionsNegativeTestJSON-376080366 tempest-InstanceActionsNegativeTestJSON-376080366-project-member] Lock "68e791b4-61db-4b6b-a30a-ccf5d6657643" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.574s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.344163] env[62183]: DEBUG oslo_vmware.api [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387070, 'name': PowerOffVM_Task, 'duration_secs': 0.15372} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.348653] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.349727] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.349727] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7658eb6c-ecd5-4e7c-84a6-9e32b97e412b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.357178] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387071, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06918} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.357589] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.358669] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0092ca4-485d-4171-9c48-a94497ed687a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.384399] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] a948464b-63aa-4bc8-9885-228049e96d37/a948464b-63aa-4bc8-9885-228049e96d37.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.386311] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4f57c0d-5991-48bd-9545-c7f21c21f5d4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.402675] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.403170] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.403700] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Deleting the datastore file [datastore2] 4fa0beb2-3f36-48f4-996c-1da0bb891f93 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.405550] env[62183]: DEBUG nova.network.neutron [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updating instance_info_cache with network_info: [{"id": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "address": "fa:16:3e:3f:c5:6f", "network": {"id": "7d52241a-8c3c-46bc-b64d-1cbc74ec2434", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-104877467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e74d26cb8f5a47a48dc025b8058c5768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fafdd3-76", "ovs_interfaceid": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.406545] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c2a26f5-ea53-411f-b751-f59ac506dc5d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.418019] env[62183]: DEBUG oslo_vmware.api [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for the task: (returnval){ [ 860.418019] env[62183]: value = "task-1387073" [ 860.418019] env[62183]: _type = "Task" [ 860.418019] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.418019] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 860.418019] env[62183]: value = "task-1387074" [ 860.418019] env[62183]: _type = "Task" [ 860.418019] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.433303] env[62183]: DEBUG oslo_vmware.api [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387073, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.437208] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387074, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.456040] env[62183]: DEBUG nova.compute.manager [req-40fffd69-6a7d-4d19-a580-2e6f4b5cdfd3 req-48a6f127-6406-41d7-a1f4-585d577cc318 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Received event network-vif-plugged-4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.456040] env[62183]: DEBUG oslo_concurrency.lockutils [req-40fffd69-6a7d-4d19-a580-2e6f4b5cdfd3 req-48a6f127-6406-41d7-a1f4-585d577cc318 service nova] Acquiring lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.456040] env[62183]: DEBUG oslo_concurrency.lockutils [req-40fffd69-6a7d-4d19-a580-2e6f4b5cdfd3 req-48a6f127-6406-41d7-a1f4-585d577cc318 service nova] Lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.456040] env[62183]: DEBUG oslo_concurrency.lockutils [req-40fffd69-6a7d-4d19-a580-2e6f4b5cdfd3 req-48a6f127-6406-41d7-a1f4-585d577cc318 service nova] Lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.456040] env[62183]: DEBUG nova.compute.manager [req-40fffd69-6a7d-4d19-a580-2e6f4b5cdfd3 req-48a6f127-6406-41d7-a1f4-585d577cc318 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] No waiting events found dispatching network-vif-plugged-4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.456423] env[62183]: WARNING nova.compute.manager [req-40fffd69-6a7d-4d19-a580-2e6f4b5cdfd3 req-48a6f127-6406-41d7-a1f4-585d577cc318 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Received unexpected event network-vif-plugged-4b70b794-46e6-48de-8932-98875906aa5a for instance with vm_state building and task_state spawning. [ 860.486946] env[62183]: DEBUG nova.network.neutron [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Successfully updated port: 4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.546341] env[62183]: DEBUG oslo_vmware.rw_handles [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52614986-bf30-669f-0691-ea649b1c59e6/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 860.547315] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e722efa-b2ae-45a5-9516-5d51ce013192 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.553023] env[62183]: DEBUG oslo_vmware.rw_handles [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52614986-bf30-669f-0691-ea649b1c59e6/disk-0.vmdk is in state: ready. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 860.553199] env[62183]: ERROR oslo_vmware.rw_handles [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52614986-bf30-669f-0691-ea649b1c59e6/disk-0.vmdk due to incomplete transfer. [ 860.556273] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e2818b89-2873-493e-8517-f894ae705974 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.557620] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387069, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.772389} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.557852] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 503786ca-dba4-43c1-9a25-9f1cbac9a6a4/503786ca-dba4-43c1-9a25-9f1cbac9a6a4.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.558094] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.558728] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9fbcffa-9aea-4f2c-bb8e-6bbc410b703a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.563391] env[62183]: DEBUG oslo_vmware.rw_handles [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52614986-bf30-669f-0691-ea649b1c59e6/disk-0.vmdk. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 860.563594] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Uploaded image 2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83 to the Glance image server {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 860.565935] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Destroying the VM {{(pid=62183) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 860.567390] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ea7976c9-f42c-4f6b-ae45-00451a1ff196 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.569063] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 860.569063] env[62183]: value = "task-1387075" [ 860.569063] env[62183]: _type = "Task" [ 860.569063] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.575430] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 860.575430] env[62183]: value = "task-1387076" [ 860.575430] env[62183]: _type = "Task" [ 860.575430] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.578575] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387075, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.585983] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387076, 'name': Destroy_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.706312] env[62183]: DEBUG nova.compute.manager [req-6cb2077b-30a6-4c79-ab60-ed658b54f1b9 req-d677bbd7-ad68-4ab2-9589-7a6e109a8b55 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received event network-changed-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.706312] env[62183]: DEBUG nova.compute.manager [req-6cb2077b-30a6-4c79-ab60-ed658b54f1b9 req-d677bbd7-ad68-4ab2-9589-7a6e109a8b55 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Refreshing instance network info cache due to event network-changed-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 860.706312] env[62183]: DEBUG oslo_concurrency.lockutils [req-6cb2077b-30a6-4c79-ab60-ed658b54f1b9 req-d677bbd7-ad68-4ab2-9589-7a6e109a8b55 service nova] Acquiring lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.911686] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.912119] env[62183]: DEBUG nova.compute.manager [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Instance network_info: |[{"id": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "address": "fa:16:3e:3f:c5:6f", "network": {"id": "7d52241a-8c3c-46bc-b64d-1cbc74ec2434", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-104877467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e74d26cb8f5a47a48dc025b8058c5768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fafdd3-76", "ovs_interfaceid": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 860.912443] env[62183]: DEBUG oslo_concurrency.lockutils [req-6cb2077b-30a6-4c79-ab60-ed658b54f1b9 req-d677bbd7-ad68-4ab2-9589-7a6e109a8b55 service nova] Acquired lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.912625] env[62183]: DEBUG nova.network.neutron [req-6cb2077b-30a6-4c79-ab60-ed658b54f1b9 req-d677bbd7-ad68-4ab2-9589-7a6e109a8b55 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Refreshing network info cache for port 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 860.913953] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:c5:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '748a5204-8f14-402c-9a6e-f3e6104db082', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37fafdd3-7651-4cfc-818d-2ea8e63ffcb1', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.921606] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Creating folder: Project (e74d26cb8f5a47a48dc025b8058c5768). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 860.927249] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-71362925-7c8c-48c6-89b3-d7e5b96a7045 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.941139] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387074, 'name': ReconfigVM_Task, 'duration_secs': 0.350226} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.946159] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Reconfigured VM instance instance-00000048 to attach disk [datastore2] a948464b-63aa-4bc8-9885-228049e96d37/a948464b-63aa-4bc8-9885-228049e96d37.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.946811] env[62183]: DEBUG oslo_vmware.api [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Task: {'id': task-1387073, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291256} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.948084] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c398128-4d82-401a-af6f-40649f9c5882 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.949694] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.949869] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.950053] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.952952] env[62183]: INFO nova.compute.manager [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Took 1.15 seconds to destroy the instance on the hypervisor. [ 860.952952] env[62183]: DEBUG oslo.service.loopingcall [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.952952] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Created folder: Project (e74d26cb8f5a47a48dc025b8058c5768) in parent group-v294392. [ 860.952952] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Creating folder: Instances. Parent ref: group-v294454. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 860.952952] env[62183]: DEBUG nova.compute.manager [-] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 860.952952] env[62183]: DEBUG nova.network.neutron [-] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 860.953181] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-95a06100-aa16-4cdb-b1cd-b8ec0be2cb0d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.958468] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 860.958468] env[62183]: value = "task-1387078" [ 860.958468] env[62183]: _type = "Task" [ 860.958468] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.965255] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Created folder: Instances in parent group-v294454. [ 860.965497] env[62183]: DEBUG oslo.service.loopingcall [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 860.965970] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 860.966208] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7b2ce67-e595-4ee7-9f80-436ad101a57f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.982905] env[62183]: DEBUG nova.network.neutron [-] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.987387] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387078, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.989058] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.989058] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquired lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.989058] env[62183]: DEBUG nova.network.neutron [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.993200] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 860.993200] env[62183]: value = "task-1387080" [ 860.993200] env[62183]: _type = "Task" [ 860.993200] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.006135] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387080, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.007842] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b862662c-4c5f-43d9-a6df-7fe08d0445dd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.014625] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1ab0bb-a48d-4ede-84d5-a13982ff95f3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.045175] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd046782-6dee-46be-b72c-26859202313b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.053391] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a17710-d38d-428a-84ee-ea16234badfa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.067607] env[62183]: DEBUG nova.compute.provider_tree [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.077202] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387075, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08263} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.078115] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.078905] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b915b1-2395-4ba5-acf5-031dfbe475c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.104535] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 503786ca-dba4-43c1-9a25-9f1cbac9a6a4/503786ca-dba4-43c1-9a25-9f1cbac9a6a4.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.110224] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c73b67c-e662-4c1a-968e-5f53d85f271a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.124050] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387076, 'name': Destroy_Task, 'duration_secs': 0.343101} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.124645] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Destroyed the VM [ 861.124984] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Deleting Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 861.125656] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d2ce81bd-c5f3-4eae-a913-7b54f9f3b14e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.130438] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 861.130438] env[62183]: value = "task-1387081" [ 861.130438] env[62183]: _type = "Task" [ 861.130438] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.134435] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 861.134435] env[62183]: value = "task-1387082" [ 861.134435] env[62183]: _type = "Task" [ 861.134435] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.142626] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.149341] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387082, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.378378] env[62183]: DEBUG nova.network.neutron [req-6cb2077b-30a6-4c79-ab60-ed658b54f1b9 req-d677bbd7-ad68-4ab2-9589-7a6e109a8b55 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updated VIF entry in instance network info cache for port 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 861.378957] env[62183]: DEBUG nova.network.neutron [req-6cb2077b-30a6-4c79-ab60-ed658b54f1b9 req-d677bbd7-ad68-4ab2-9589-7a6e109a8b55 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updating instance_info_cache with network_info: [{"id": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "address": "fa:16:3e:3f:c5:6f", "network": {"id": "7d52241a-8c3c-46bc-b64d-1cbc74ec2434", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-104877467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e74d26cb8f5a47a48dc025b8058c5768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fafdd3-76", "ovs_interfaceid": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.476126] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387078, 'name': Rename_Task, 'duration_secs': 0.147371} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.476447] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.477257] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15a072ac-7c6f-47d1-8677-384085016ccb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.484547] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 861.484547] env[62183]: value = "task-1387083" [ 861.484547] env[62183]: _type = "Task" [ 861.484547] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.488153] env[62183]: DEBUG nova.network.neutron [-] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.494484] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.504983] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387080, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.561474] env[62183]: DEBUG nova.network.neutron [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.570809] env[62183]: DEBUG nova.scheduler.client.report [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 861.642435] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387081, 'name': ReconfigVM_Task, 'duration_secs': 0.350986} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.643274] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 503786ca-dba4-43c1-9a25-9f1cbac9a6a4/503786ca-dba4-43c1-9a25-9f1cbac9a6a4.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.643652] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83a12d41-b021-4553-adc7-f832d75bccd9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.648274] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387082, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.657335] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 861.657335] env[62183]: value = "task-1387084" [ 861.657335] env[62183]: _type = "Task" [ 861.657335] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.666453] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387084, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.740852] env[62183]: DEBUG nova.network.neutron [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Updating instance_info_cache with network_info: [{"id": "4b70b794-46e6-48de-8932-98875906aa5a", "address": "fa:16:3e:3a:a3:a7", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b70b794-46", "ovs_interfaceid": "4b70b794-46e6-48de-8932-98875906aa5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.884801] env[62183]: DEBUG oslo_concurrency.lockutils [req-6cb2077b-30a6-4c79-ab60-ed658b54f1b9 req-d677bbd7-ad68-4ab2-9589-7a6e109a8b55 service nova] Releasing lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.985992] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "be021940-5fff-429a-9b93-4ad238cdff77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.986353] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "be021940-5fff-429a-9b93-4ad238cdff77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.990924] env[62183]: INFO nova.compute.manager [-] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Took 1.04 seconds to deallocate network for instance. [ 862.002443] env[62183]: DEBUG oslo_vmware.api [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387083, 'name': PowerOnVM_Task, 'duration_secs': 0.510212} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.003358] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.003951] env[62183]: INFO nova.compute.manager [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Took 9.04 seconds to spawn the instance on the hypervisor. [ 862.003951] env[62183]: DEBUG nova.compute.manager [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.004888] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f218a2-07b9-4766-bcc2-419672f2eb45 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.010916] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387080, 'name': CreateVM_Task, 'duration_secs': 0.792266} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.011500] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.012112] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.012288] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.012608] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.013095] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89fd4be3-cb40-43ac-9df9-c23f58799fc6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.021735] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 862.021735] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52c12827-d745-df3f-5631-bf28f7bdc21b" [ 862.021735] env[62183]: _type = "Task" [ 862.021735] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.033423] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c12827-d745-df3f-5631-bf28f7bdc21b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.079407] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.080328] env[62183]: DEBUG nova.compute.manager [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 862.082759] env[62183]: DEBUG oslo_concurrency.lockutils [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.067s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.082990] env[62183]: DEBUG nova.objects.instance [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lazy-loading 'resources' on Instance uuid 39683931-d0ff-4a5c-a4a2-792230ab0e3d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.144615] env[62183]: DEBUG oslo_vmware.api [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387082, 'name': RemoveSnapshot_Task, 'duration_secs': 0.718431} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.144884] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Deleted Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 862.145143] env[62183]: INFO nova.compute.manager [None req-3a47116e-7f2c-4ce9-8582-83a1784fa06c tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Took 15.11 seconds to snapshot the instance on the hypervisor. [ 862.164658] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387084, 'name': Rename_Task, 'duration_secs': 0.170153} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.165278] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.165278] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de0ab394-5afa-4500-a812-6b00976f16a3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.171177] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 862.171177] env[62183]: value = "task-1387085" [ 862.171177] env[62183]: _type = "Task" [ 862.171177] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.184219] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387085, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.243693] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Releasing lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.243854] env[62183]: DEBUG nova.compute.manager [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Instance network_info: |[{"id": "4b70b794-46e6-48de-8932-98875906aa5a", "address": "fa:16:3e:3a:a3:a7", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b70b794-46", "ovs_interfaceid": "4b70b794-46e6-48de-8932-98875906aa5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.244407] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:a3:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0721b358-3768-472d-95f8-6d6755ab1635', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b70b794-46e6-48de-8932-98875906aa5a', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.253276] env[62183]: DEBUG oslo.service.loopingcall [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.253646] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.253906] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b690183-9e63-46ff-9fc5-6c774b0b0959 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.277422] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.277422] env[62183]: value = "task-1387086" [ 862.277422] env[62183]: _type = "Task" [ 862.277422] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.286555] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387086, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.480558] env[62183]: DEBUG nova.compute.manager [req-61d5c86c-f7cc-460a-9a47-dc7e3b8e662e req-976ac49d-db57-47f9-8f33-795adce08ef2 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Received event network-changed-4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.480850] env[62183]: DEBUG nova.compute.manager [req-61d5c86c-f7cc-460a-9a47-dc7e3b8e662e req-976ac49d-db57-47f9-8f33-795adce08ef2 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Refreshing instance network info cache due to event network-changed-4b70b794-46e6-48de-8932-98875906aa5a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.481149] env[62183]: DEBUG oslo_concurrency.lockutils [req-61d5c86c-f7cc-460a-9a47-dc7e3b8e662e req-976ac49d-db57-47f9-8f33-795adce08ef2 service nova] Acquiring lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.481344] env[62183]: DEBUG oslo_concurrency.lockutils [req-61d5c86c-f7cc-460a-9a47-dc7e3b8e662e req-976ac49d-db57-47f9-8f33-795adce08ef2 service nova] Acquired lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.481535] env[62183]: DEBUG nova.network.neutron [req-61d5c86c-f7cc-460a-9a47-dc7e3b8e662e req-976ac49d-db57-47f9-8f33-795adce08ef2 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Refreshing network info cache for port 4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.490720] env[62183]: DEBUG nova.compute.manager [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.505406] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.534843] env[62183]: INFO nova.compute.manager [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Took 24.14 seconds to build instance. [ 862.542262] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c12827-d745-df3f-5631-bf28f7bdc21b, 'name': SearchDatastore_Task, 'duration_secs': 0.023107} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.542580] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.542813] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.543099] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.543215] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.543399] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.543664] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3d789c3-38f2-4efa-a4e8-c8dca6e8eab3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.552442] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.552610] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 862.553864] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40182b65-fabf-4483-84ef-01cda406d15a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.559291] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 862.559291] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]527eecbe-34a6-e87c-20f5-43e593d849df" [ 862.559291] env[62183]: _type = "Task" [ 862.559291] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.567327] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527eecbe-34a6-e87c-20f5-43e593d849df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.587864] env[62183]: DEBUG nova.compute.utils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.593898] env[62183]: DEBUG nova.compute.manager [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.594273] env[62183]: DEBUG nova.network.neutron [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 862.663362] env[62183]: DEBUG nova.policy [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e0eee67c7954c80b4e1ea43b0abbcb5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebcc716f958942b588a6bfde78d2c00d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.687110] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387085, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.794974] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387086, 'name': CreateVM_Task, 'duration_secs': 0.481074} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.795095] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.795996] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.796179] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.796496] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.796756] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dfab2be-d0e4-4939-ba88-d0b9301a1196 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.806705] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 862.806705] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e27a46-318d-b78a-ff64-2cae711d8658" [ 862.806705] env[62183]: _type = "Task" [ 862.806705] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.813657] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e27a46-318d-b78a-ff64-2cae711d8658, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.829231] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe9d79c-75ef-4d67-a8fb-896147017014 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.834305] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc09932-629f-40ea-b50c-add1dd8ca56b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.868048] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9483efeb-4540-46ba-9e6c-925070e84c02 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.875953] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092352d0-7793-4796-b5ee-8ead7932a801 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.889230] env[62183]: DEBUG nova.compute.provider_tree [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.014598] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.037959] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4a9bf7b3-9f39-4460-a5b8-013a4c8a5c1f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.383s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.069682] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527eecbe-34a6-e87c-20f5-43e593d849df, 'name': SearchDatastore_Task, 'duration_secs': 0.016216} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.071191] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffbaff2e-ca22-4441-a8c4-75e5e4919e26 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.077012] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 863.077012] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52fe95b7-9a11-ff72-d29c-cc59f94fe989" [ 863.077012] env[62183]: _type = "Task" [ 863.077012] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.085604] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52fe95b7-9a11-ff72-d29c-cc59f94fe989, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.094689] env[62183]: DEBUG nova.compute.manager [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.186273] env[62183]: DEBUG oslo_vmware.api [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387085, 'name': PowerOnVM_Task, 'duration_secs': 0.697556} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.186515] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.186722] env[62183]: INFO nova.compute.manager [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Took 7.89 seconds to spawn the instance on the hypervisor. [ 863.187112] env[62183]: DEBUG nova.compute.manager [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.187686] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91903c95-651d-48bf-a8de-5a132b6652e9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.197657] env[62183]: DEBUG nova.network.neutron [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Successfully created port: 999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.318958] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e27a46-318d-b78a-ff64-2cae711d8658, 'name': SearchDatastore_Task, 'duration_secs': 0.017499} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.319365] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.319722] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.319996] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.328670] env[62183]: DEBUG nova.network.neutron [req-61d5c86c-f7cc-460a-9a47-dc7e3b8e662e req-976ac49d-db57-47f9-8f33-795adce08ef2 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Updated VIF entry in instance network info cache for port 4b70b794-46e6-48de-8932-98875906aa5a. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.329042] env[62183]: DEBUG nova.network.neutron [req-61d5c86c-f7cc-460a-9a47-dc7e3b8e662e req-976ac49d-db57-47f9-8f33-795adce08ef2 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Updating instance_info_cache with network_info: [{"id": "4b70b794-46e6-48de-8932-98875906aa5a", "address": "fa:16:3e:3a:a3:a7", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b70b794-46", "ovs_interfaceid": "4b70b794-46e6-48de-8932-98875906aa5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.392619] env[62183]: DEBUG nova.scheduler.client.report [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 863.587741] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52fe95b7-9a11-ff72-d29c-cc59f94fe989, 'name': SearchDatastore_Task, 'duration_secs': 0.021509} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.588093] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.588378] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9/fb348784-62a0-4d1f-ac7f-f176f3da0dd9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 863.588702] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.588937] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.589385] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-11e83189-32ca-4009-bcc8-f88f0eedfee5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.591161] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-601501bb-8635-4a78-86e6-861f236541c9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.601619] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 863.601619] env[62183]: value = "task-1387087" [ 863.601619] env[62183]: _type = "Task" [ 863.601619] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.603036] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.603262] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.606987] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c748b516-c75a-40cf-ad52-4cb9c604e851 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.614171] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387087, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.615637] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 863.615637] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a8817e-93ae-76e4-033f-27aaee9491f2" [ 863.615637] env[62183]: _type = "Task" [ 863.615637] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.624250] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a8817e-93ae-76e4-033f-27aaee9491f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.710725] env[62183]: INFO nova.compute.manager [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Took 22.83 seconds to build instance. [ 863.832137] env[62183]: DEBUG oslo_concurrency.lockutils [req-61d5c86c-f7cc-460a-9a47-dc7e3b8e662e req-976ac49d-db57-47f9-8f33-795adce08ef2 service nova] Releasing lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.897931] env[62183]: DEBUG oslo_concurrency.lockutils [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.815s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.900452] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.616s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.902132] env[62183]: INFO nova.compute.claims [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.931378] env[62183]: INFO nova.scheduler.client.report [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Deleted allocations for instance 39683931-d0ff-4a5c-a4a2-792230ab0e3d [ 864.103690] env[62183]: DEBUG nova.compute.manager [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 864.116180] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387087, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.126418] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a8817e-93ae-76e4-033f-27aaee9491f2, 'name': SearchDatastore_Task, 'duration_secs': 0.017612} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.131954] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c44e7944-686a-4608-b8a4-e2b6b74218cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.135931] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 864.135931] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52c8eb23-4a15-ed02-e25f-a16e876f0e28" [ 864.135931] env[62183]: _type = "Task" [ 864.135931] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.142718] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.143058] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.143272] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.143505] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.143696] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.143904] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.144201] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.144412] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.144648] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.144887] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.145177] env[62183]: DEBUG nova.virt.hardware [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.146196] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290d8213-354b-4e42-b8ac-3a9183ffca0d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.156581] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c8eb23-4a15-ed02-e25f-a16e876f0e28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.160197] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d6deb1-8198-41db-9ba7-192dcef43691 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.212975] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b9f69940-1834-4589-a6d2-36f1c2dfa2b7 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.645s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.446500] env[62183]: DEBUG oslo_concurrency.lockutils [None req-59ea6563-6e02-42a5-bd53-5281a5aa0a44 tempest-AttachInterfacesV270Test-1984451851 tempest-AttachInterfacesV270Test-1984451851-project-member] Lock "39683931-d0ff-4a5c-a4a2-792230ab0e3d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.994s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.519973] env[62183]: DEBUG nova.compute.manager [req-2ba19e83-a930-4246-b0c8-49d3f59abca1 req-635d2ae7-37b2-4e41-8535-8596ae6304bf service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Received event network-changed-3d687443-41fd-44ff-8981-35319b79d88a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.520205] env[62183]: DEBUG nova.compute.manager [req-2ba19e83-a930-4246-b0c8-49d3f59abca1 req-635d2ae7-37b2-4e41-8535-8596ae6304bf service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Refreshing instance network info cache due to event network-changed-3d687443-41fd-44ff-8981-35319b79d88a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.520421] env[62183]: DEBUG oslo_concurrency.lockutils [req-2ba19e83-a930-4246-b0c8-49d3f59abca1 req-635d2ae7-37b2-4e41-8535-8596ae6304bf service nova] Acquiring lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.520558] env[62183]: DEBUG oslo_concurrency.lockutils [req-2ba19e83-a930-4246-b0c8-49d3f59abca1 req-635d2ae7-37b2-4e41-8535-8596ae6304bf service nova] Acquired lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.521159] env[62183]: DEBUG nova.network.neutron [req-2ba19e83-a930-4246-b0c8-49d3f59abca1 req-635d2ae7-37b2-4e41-8535-8596ae6304bf service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Refreshing network info cache for port 3d687443-41fd-44ff-8981-35319b79d88a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 864.621342] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387087, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.895212} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.621342] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9/fb348784-62a0-4d1f-ac7f-f176f3da0dd9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 864.621342] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.621342] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0742c994-c070-40f9-b197-5020234b8489 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.628406] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 864.628406] env[62183]: value = "task-1387088" [ 864.628406] env[62183]: _type = "Task" [ 864.628406] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.643068] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387088, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.651015] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c8eb23-4a15-ed02-e25f-a16e876f0e28, 'name': SearchDatastore_Task, 'duration_secs': 0.05788} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.651514] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.651514] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 7b5f50f8-5527-4f29-9fa6-6a598b77ef80/7b5f50f8-5527-4f29-9fa6-6a598b77ef80.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.651828] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f12b790b-15d6-4070-9ec0-b9e5d4d109d7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.659475] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 864.659475] env[62183]: value = "task-1387089" [ 864.659475] env[62183]: _type = "Task" [ 864.659475] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.668497] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387089, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.144480] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387088, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.159041} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.144480] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.144480] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a0c6b7-8007-462b-bbd4-276d6babb537 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.170196] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9/fb348784-62a0-4d1f-ac7f-f176f3da0dd9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.174368] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6b8f935-ddb5-4af1-b802-e446ddf66f87 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.192251] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb634b8-d48f-421e-b73b-56fede90df5a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.200774] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387089, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478818} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.203415] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 7b5f50f8-5527-4f29-9fa6-6a598b77ef80/7b5f50f8-5527-4f29-9fa6-6a598b77ef80.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 865.203560] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.203771] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 865.203771] env[62183]: value = "task-1387090" [ 865.203771] env[62183]: _type = "Task" [ 865.203771] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.204055] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09ad9c6d-31bd-4b5d-a8b2-950e7671cd26 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.206784] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdd4ac3-a8d4-4bc6-b744-6d0d5e4f1a24 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.219618] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387090, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.245831] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 865.245831] env[62183]: value = "task-1387091" [ 865.245831] env[62183]: _type = "Task" [ 865.245831] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.249591] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d9168c-e570-4d92-8494-8d11f1bfdedc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.262939] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e67a059-10f2-4708-842d-66a8ea27d3fb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.267188] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387091, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.276923] env[62183]: DEBUG nova.compute.provider_tree [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.319285] env[62183]: DEBUG nova.network.neutron [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Successfully updated port: 999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.716409] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387090, 'name': ReconfigVM_Task, 'duration_secs': 0.280629} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.716686] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Reconfigured VM instance instance-0000004a to attach disk [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9/fb348784-62a0-4d1f-ac7f-f176f3da0dd9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.717350] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17501bcd-f9b1-4842-8864-83a526a19b97 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.724902] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 865.724902] env[62183]: value = "task-1387092" [ 865.724902] env[62183]: _type = "Task" [ 865.724902] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.739521] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387092, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.739521] env[62183]: DEBUG nova.network.neutron [req-2ba19e83-a930-4246-b0c8-49d3f59abca1 req-635d2ae7-37b2-4e41-8535-8596ae6304bf service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updated VIF entry in instance network info cache for port 3d687443-41fd-44ff-8981-35319b79d88a. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.739591] env[62183]: DEBUG nova.network.neutron [req-2ba19e83-a930-4246-b0c8-49d3f59abca1 req-635d2ae7-37b2-4e41-8535-8596ae6304bf service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [{"id": "3d687443-41fd-44ff-8981-35319b79d88a", "address": "fa:16:3e:91:31:9b", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d687443-41", "ovs_interfaceid": "3d687443-41fd-44ff-8981-35319b79d88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.757296] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387091, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.759388] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.759508] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec62a38-2f6c-4363-840e-f02ce1f6a8d2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.789305] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 7b5f50f8-5527-4f29-9fa6-6a598b77ef80/7b5f50f8-5527-4f29-9fa6-6a598b77ef80.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.790336] env[62183]: DEBUG nova.scheduler.client.report [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 865.794595] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-826ba14b-735c-4401-b46f-9f997bdb73e3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.815983] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 865.815983] env[62183]: value = "task-1387093" [ 865.815983] env[62183]: _type = "Task" [ 865.815983] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.826885] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.827058] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.827217] env[62183]: DEBUG nova.network.neutron [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.829157] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387093, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.243902] env[62183]: DEBUG oslo_concurrency.lockutils [req-2ba19e83-a930-4246-b0c8-49d3f59abca1 req-635d2ae7-37b2-4e41-8535-8596ae6304bf service nova] Releasing lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.244405] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387092, 'name': Rename_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.310017] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.310617] env[62183]: DEBUG nova.compute.manager [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.313661] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 7.796s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.313856] env[62183]: DEBUG nova.objects.instance [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62183) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 866.328073] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387093, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.383177] env[62183]: DEBUG nova.network.neutron [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.551601] env[62183]: DEBUG nova.compute.manager [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received event network-vif-plugged-999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.551897] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] Acquiring lock "1b0593f2-b712-4926-bfb8-30c92070ff6a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.552890] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.552890] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.553051] env[62183]: DEBUG nova.compute.manager [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] No waiting events found dispatching network-vif-plugged-999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 866.553095] env[62183]: WARNING nova.compute.manager [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received unexpected event network-vif-plugged-999bf20b-d0e4-4229-a150-ec7c31e38cc6 for instance with vm_state building and task_state spawning. [ 866.553255] env[62183]: DEBUG nova.compute.manager [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received event network-changed-999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.553438] env[62183]: DEBUG nova.compute.manager [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing instance network info cache due to event network-changed-999bf20b-d0e4-4229-a150-ec7c31e38cc6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 866.553584] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] Acquiring lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.701111] env[62183]: DEBUG nova.network.neutron [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updating instance_info_cache with network_info: [{"id": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "address": "fa:16:3e:46:24:80", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap999bf20b-d0", "ovs_interfaceid": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.739909] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387092, 'name': Rename_Task, 'duration_secs': 0.886932} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.739909] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.739909] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1d8bcca-0f43-4649-b385-73bfa9c0b283 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.746155] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 866.746155] env[62183]: value = "task-1387094" [ 866.746155] env[62183]: _type = "Task" [ 866.746155] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.756747] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387094, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.822841] env[62183]: DEBUG nova.compute.utils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.826018] env[62183]: DEBUG nova.compute.manager [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Stashing vm_state: active {{(pid=62183) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 866.828594] env[62183]: DEBUG nova.compute.manager [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 866.828594] env[62183]: DEBUG nova.network.neutron [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 866.843079] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387093, 'name': ReconfigVM_Task, 'duration_secs': 0.92279} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.843608] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 7b5f50f8-5527-4f29-9fa6-6a598b77ef80/7b5f50f8-5527-4f29-9fa6-6a598b77ef80.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.845150] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b1ab2112-aae9-4a72-b43f-b21cdef0b929 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.852659] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 866.852659] env[62183]: value = "task-1387095" [ 866.852659] env[62183]: _type = "Task" [ 866.852659] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.861706] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387095, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.907914] env[62183]: DEBUG nova.policy [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59edae5aed2e4ccbb980bc0973890baf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '594f9c41496146c68921dad4c8d99ec4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.207504] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.207978] env[62183]: DEBUG nova.compute.manager [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Instance network_info: |[{"id": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "address": "fa:16:3e:46:24:80", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap999bf20b-d0", "ovs_interfaceid": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 867.208405] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] Acquired lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.208647] env[62183]: DEBUG nova.network.neutron [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing network info cache for port 999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.210293] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:24:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db00ec2e-3155-46b6-8170-082f7d86dbe7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '999bf20b-d0e4-4229-a150-ec7c31e38cc6', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.219595] env[62183]: DEBUG oslo.service.loopingcall [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.226441] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.226441] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae0d945d-0265-4b7e-843f-d37a251d5442 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.251541] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.251541] env[62183]: value = "task-1387096" [ 867.251541] env[62183]: _type = "Task" [ 867.251541] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.266665] env[62183]: DEBUG oslo_vmware.api [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387094, 'name': PowerOnVM_Task, 'duration_secs': 0.505891} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.272539] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.272539] env[62183]: INFO nova.compute.manager [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Took 9.67 seconds to spawn the instance on the hypervisor. [ 867.272942] env[62183]: DEBUG nova.compute.manager [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.273160] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387096, 'name': CreateVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.274783] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c4373c-5b3b-45f4-b5d1-dfd4221c526b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.330242] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1d837c41-8769-40ee-b63d-8c09ecc06d9d tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.330735] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.825s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.330987] env[62183]: DEBUG nova.objects.instance [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lazy-loading 'resources' on Instance uuid 4fa0beb2-3f36-48f4-996c-1da0bb891f93 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.336752] env[62183]: DEBUG nova.compute.manager [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.362842] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.368404] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387095, 'name': Rename_Task, 'duration_secs': 0.161326} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.368524] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.369063] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-836ab0f9-db86-4e63-bddc-261ba81ae333 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.377381] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 867.377381] env[62183]: value = "task-1387097" [ 867.377381] env[62183]: _type = "Task" [ 867.377381] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.393360] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387097, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.450160] env[62183]: DEBUG nova.network.neutron [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Successfully created port: 4266b7f9-08bc-469d-b4e8-a88b751a164e {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.610345] env[62183]: DEBUG nova.network.neutron [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updated VIF entry in instance network info cache for port 999bf20b-d0e4-4229-a150-ec7c31e38cc6. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 867.610987] env[62183]: DEBUG nova.network.neutron [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updating instance_info_cache with network_info: [{"id": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "address": "fa:16:3e:46:24:80", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap999bf20b-d0", "ovs_interfaceid": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.768066] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387096, 'name': CreateVM_Task, 'duration_secs': 0.418434} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.768066] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.768066] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.768066] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.768994] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.768994] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c7b4bac-7a4e-4cb4-8179-3befeb5ec536 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.773772] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 867.773772] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]520492f9-74c1-d8e7-74cf-5f6ac26f668b" [ 867.773772] env[62183]: _type = "Task" [ 867.773772] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.784771] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]520492f9-74c1-d8e7-74cf-5f6ac26f668b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.796790] env[62183]: INFO nova.compute.manager [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Took 26.09 seconds to build instance. [ 867.828787] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.829029] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.895469] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387097, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.070668] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48b0951-0d19-4e9d-8638-0fd3ad230881 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.078784] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bf34d6-a681-4aaf-ac6b-52ba9e1d320f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.109390] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a882c2dd-ae95-4ff1-a961-11c872a6eec6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.113851] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7dc3975-0ae0-4c2b-9ed6-ff0c98a05bc9 req-03e4dffb-ea0d-4f8c-b696-f9bffc1a11e3 service nova] Releasing lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.117353] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9033be-d1a5-433f-9ae8-7bd041e5b0a6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.133062] env[62183]: DEBUG nova.compute.provider_tree [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.288642] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]520492f9-74c1-d8e7-74cf-5f6ac26f668b, 'name': SearchDatastore_Task, 'duration_secs': 0.011724} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.291144] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.291144] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.291144] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.291144] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.291384] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.291550] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b2cb1fb-9d56-4edb-8b6f-a975abc6add6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.298738] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3c08ccbc-bbcc-4b1e-8e67-598bbb2285d7 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.590s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.300921] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.301135] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 868.301908] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7c6b222-b814-4728-ac4c-21798955bb6f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.307702] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 868.307702] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5283635c-9ff3-b2db-c8e7-f188043d803c" [ 868.307702] env[62183]: _type = "Task" [ 868.307702] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.316298] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5283635c-9ff3-b2db-c8e7-f188043d803c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.331893] env[62183]: DEBUG nova.compute.manager [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 868.349293] env[62183]: DEBUG nova.compute.manager [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 868.381099] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.381418] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.381601] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.381818] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.381988] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.382267] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.382522] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.382735] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.382943] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.383160] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.383385] env[62183]: DEBUG nova.virt.hardware [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.384550] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89aea42-a74c-45c4-970f-b45f6575a754 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.399417] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186838c3-a44f-4f2b-b172-fe273eb057d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.404985] env[62183]: DEBUG oslo_vmware.api [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387097, 'name': PowerOnVM_Task, 'duration_secs': 0.750858} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.405749] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.405837] env[62183]: INFO nova.compute.manager [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Took 8.48 seconds to spawn the instance on the hypervisor. [ 868.406141] env[62183]: DEBUG nova.compute.manager [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.407416] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9cabd05-b0fa-4825-88eb-9a0627169fee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.637347] env[62183]: DEBUG nova.scheduler.client.report [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 868.675726] env[62183]: DEBUG nova.compute.manager [req-26b85da5-ff4c-4ca4-8cd2-5dd4efd6dd90 req-0945e1d4-ead6-4cf9-bf5e-7366f12530d0 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received event network-changed-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.675933] env[62183]: DEBUG nova.compute.manager [req-26b85da5-ff4c-4ca4-8cd2-5dd4efd6dd90 req-0945e1d4-ead6-4cf9-bf5e-7366f12530d0 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Refreshing instance network info cache due to event network-changed-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 868.676178] env[62183]: DEBUG oslo_concurrency.lockutils [req-26b85da5-ff4c-4ca4-8cd2-5dd4efd6dd90 req-0945e1d4-ead6-4cf9-bf5e-7366f12530d0 service nova] Acquiring lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.676330] env[62183]: DEBUG oslo_concurrency.lockutils [req-26b85da5-ff4c-4ca4-8cd2-5dd4efd6dd90 req-0945e1d4-ead6-4cf9-bf5e-7366f12530d0 service nova] Acquired lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.676492] env[62183]: DEBUG nova.network.neutron [req-26b85da5-ff4c-4ca4-8cd2-5dd4efd6dd90 req-0945e1d4-ead6-4cf9-bf5e-7366f12530d0 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Refreshing network info cache for port 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 868.822905] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5283635c-9ff3-b2db-c8e7-f188043d803c, 'name': SearchDatastore_Task, 'duration_secs': 0.011438} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.824193] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28cd4a80-8584-452d-9c27-3cdb06d7b5ed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.830264] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 868.830264] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52961853-60d4-fc42-294c-68349a15b594" [ 868.830264] env[62183]: _type = "Task" [ 868.830264] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.845636] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52961853-60d4-fc42-294c-68349a15b594, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.868928] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.939614] env[62183]: INFO nova.compute.manager [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Took 25.06 seconds to build instance. [ 869.142615] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.812s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.146805] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.131s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.150167] env[62183]: INFO nova.compute.claims [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.177062] env[62183]: INFO nova.scheduler.client.report [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Deleted allocations for instance 4fa0beb2-3f36-48f4-996c-1da0bb891f93 [ 869.239725] env[62183]: DEBUG nova.network.neutron [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Successfully updated port: 4266b7f9-08bc-469d-b4e8-a88b751a164e {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.340574] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52961853-60d4-fc42-294c-68349a15b594, 'name': SearchDatastore_Task, 'duration_secs': 0.029168} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.340873] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.341134] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 1b0593f2-b712-4926-bfb8-30c92070ff6a/1b0593f2-b712-4926-bfb8-30c92070ff6a.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 869.341411] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bee7011e-e256-4ddf-97a3-25ecabab99cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.348591] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 869.348591] env[62183]: value = "task-1387098" [ 869.348591] env[62183]: _type = "Task" [ 869.348591] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.356028] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387098, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.443563] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e5f1ce1-ca3e-4b2d-878c-3284de282716 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.656s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.694401] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9d3adb73-8604-4ba5-8bcf-e866ffa814ca tempest-ServerShowV257Test-117959780 tempest-ServerShowV257Test-117959780-project-member] Lock "4fa0beb2-3f36-48f4-996c-1da0bb891f93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.006s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.741958] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-843d35ff-6871-4e2a-8e03-d7229a0d8246" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.742257] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-843d35ff-6871-4e2a-8e03-d7229a0d8246" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.742433] env[62183]: DEBUG nova.network.neutron [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.858722] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387098, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.988125] env[62183]: DEBUG nova.network.neutron [req-26b85da5-ff4c-4ca4-8cd2-5dd4efd6dd90 req-0945e1d4-ead6-4cf9-bf5e-7366f12530d0 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updated VIF entry in instance network info cache for port 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 869.988125] env[62183]: DEBUG nova.network.neutron [req-26b85da5-ff4c-4ca4-8cd2-5dd4efd6dd90 req-0945e1d4-ead6-4cf9-bf5e-7366f12530d0 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updating instance_info_cache with network_info: [{"id": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "address": "fa:16:3e:3f:c5:6f", "network": {"id": "7d52241a-8c3c-46bc-b64d-1cbc74ec2434", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-104877467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e74d26cb8f5a47a48dc025b8058c5768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fafdd3-76", "ovs_interfaceid": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.016347] env[62183]: DEBUG nova.compute.manager [req-c9f60f33-3f33-4c21-8b8c-d7ad0deb3464 req-f800dc03-1839-4df5-a2b9-c831d651e36c service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Received event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.017332] env[62183]: DEBUG nova.compute.manager [req-c9f60f33-3f33-4c21-8b8c-d7ad0deb3464 req-f800dc03-1839-4df5-a2b9-c831d651e36c service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing instance network info cache due to event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.017332] env[62183]: DEBUG oslo_concurrency.lockutils [req-c9f60f33-3f33-4c21-8b8c-d7ad0deb3464 req-f800dc03-1839-4df5-a2b9-c831d651e36c service nova] Acquiring lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.017332] env[62183]: DEBUG oslo_concurrency.lockutils [req-c9f60f33-3f33-4c21-8b8c-d7ad0deb3464 req-f800dc03-1839-4df5-a2b9-c831d651e36c service nova] Acquired lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.017332] env[62183]: DEBUG nova.network.neutron [req-c9f60f33-3f33-4c21-8b8c-d7ad0deb3464 req-f800dc03-1839-4df5-a2b9-c831d651e36c service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing network info cache for port dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.287256] env[62183]: DEBUG nova.network.neutron [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.360609] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387098, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606582} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.363344] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 1b0593f2-b712-4926-bfb8-30c92070ff6a/1b0593f2-b712-4926-bfb8-30c92070ff6a.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.363760] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.365425] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fdf0fba-963e-42e1-9745-139732b36a27 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.376222] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 870.376222] env[62183]: value = "task-1387099" [ 870.376222] env[62183]: _type = "Task" [ 870.376222] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.394182] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387099, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.453295] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2676ce-b43a-4c73-b9cb-317dc21c5e3c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.461124] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc76658-65cb-4599-8149-546eb552da46 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.494669] env[62183]: DEBUG oslo_concurrency.lockutils [req-26b85da5-ff4c-4ca4-8cd2-5dd4efd6dd90 req-0945e1d4-ead6-4cf9-bf5e-7366f12530d0 service nova] Releasing lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.495865] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4196bd0a-75e4-49ab-9afb-b96cdf861ac0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.503525] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08897f56-03ea-4c0b-ab94-1332cd3dfb0c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.509107] env[62183]: DEBUG nova.network.neutron [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Updating instance_info_cache with network_info: [{"id": "4266b7f9-08bc-469d-b4e8-a88b751a164e", "address": "fa:16:3e:13:48:39", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4266b7f9-08", "ovs_interfaceid": "4266b7f9-08bc-469d-b4e8-a88b751a164e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.519419] env[62183]: DEBUG nova.compute.provider_tree [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.806723] env[62183]: DEBUG nova.network.neutron [req-c9f60f33-3f33-4c21-8b8c-d7ad0deb3464 req-f800dc03-1839-4df5-a2b9-c831d651e36c service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updated VIF entry in instance network info cache for port dccdd9b2-d8fa-435b-86d8-ec505124174c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 870.807132] env[62183]: DEBUG nova.network.neutron [req-c9f60f33-3f33-4c21-8b8c-d7ad0deb3464 req-f800dc03-1839-4df5-a2b9-c831d651e36c service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updating instance_info_cache with network_info: [{"id": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "address": "fa:16:3e:f8:f1:0b", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdccdd9b2-d8", "ovs_interfaceid": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.867237] env[62183]: DEBUG nova.compute.manager [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Received event network-vif-plugged-4266b7f9-08bc-469d-b4e8-a88b751a164e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.867461] env[62183]: DEBUG oslo_concurrency.lockutils [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] Acquiring lock "843d35ff-6871-4e2a-8e03-d7229a0d8246-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.867689] env[62183]: DEBUG oslo_concurrency.lockutils [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.867870] env[62183]: DEBUG oslo_concurrency.lockutils [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.867986] env[62183]: DEBUG nova.compute.manager [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] No waiting events found dispatching network-vif-plugged-4266b7f9-08bc-469d-b4e8-a88b751a164e {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.868275] env[62183]: WARNING nova.compute.manager [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Received unexpected event network-vif-plugged-4266b7f9-08bc-469d-b4e8-a88b751a164e for instance with vm_state building and task_state spawning. [ 870.868451] env[62183]: DEBUG nova.compute.manager [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Received event network-changed-4266b7f9-08bc-469d-b4e8-a88b751a164e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.868609] env[62183]: DEBUG nova.compute.manager [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Refreshing instance network info cache due to event network-changed-4266b7f9-08bc-469d-b4e8-a88b751a164e. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.868866] env[62183]: DEBUG oslo_concurrency.lockutils [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] Acquiring lock "refresh_cache-843d35ff-6871-4e2a-8e03-d7229a0d8246" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.892218] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387099, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077031} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.892627] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.893715] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7189e0b5-6f5e-4164-afb3-5dcfd9165b79 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.922081] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 1b0593f2-b712-4926-bfb8-30c92070ff6a/1b0593f2-b712-4926-bfb8-30c92070ff6a.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.922396] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbc5e7d5-131a-4da7-84e0-5a2248cbf3e4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.946994] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 870.946994] env[62183]: value = "task-1387100" [ 870.946994] env[62183]: _type = "Task" [ 870.946994] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.957724] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387100, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.021618] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-843d35ff-6871-4e2a-8e03-d7229a0d8246" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.022057] env[62183]: DEBUG nova.compute.manager [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Instance network_info: |[{"id": "4266b7f9-08bc-469d-b4e8-a88b751a164e", "address": "fa:16:3e:13:48:39", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4266b7f9-08", "ovs_interfaceid": "4266b7f9-08bc-469d-b4e8-a88b751a164e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 871.023168] env[62183]: DEBUG nova.scheduler.client.report [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 871.026832] env[62183]: DEBUG oslo_concurrency.lockutils [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] Acquired lock "refresh_cache-843d35ff-6871-4e2a-8e03-d7229a0d8246" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.027088] env[62183]: DEBUG nova.network.neutron [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Refreshing network info cache for port 4266b7f9-08bc-469d-b4e8-a88b751a164e {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.028564] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:48:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '496ac502-bfc4-4324-8332-cac473eb7cc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4266b7f9-08bc-469d-b4e8-a88b751a164e', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 871.037376] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Creating folder: Project (594f9c41496146c68921dad4c8d99ec4). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 871.041171] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dba80494-5dc9-4f29-bb74-605e44c82250 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.055413] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Created folder: Project (594f9c41496146c68921dad4c8d99ec4) in parent group-v294392. [ 871.055679] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Creating folder: Instances. Parent ref: group-v294459. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 871.055957] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-10b59eb8-613e-43b3-aa8b-1e565044624a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.065432] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Created folder: Instances in parent group-v294459. [ 871.065715] env[62183]: DEBUG oslo.service.loopingcall [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.065915] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 871.066279] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b2e31b4-0ebd-4407-9e24-8b3f2b265fd5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.095307] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 871.095307] env[62183]: value = "task-1387103" [ 871.095307] env[62183]: _type = "Task" [ 871.095307] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.105933] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387103, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.310675] env[62183]: DEBUG oslo_concurrency.lockutils [req-c9f60f33-3f33-4c21-8b8c-d7ad0deb3464 req-f800dc03-1839-4df5-a2b9-c831d651e36c service nova] Releasing lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.439033] env[62183]: DEBUG nova.network.neutron [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Updated VIF entry in instance network info cache for port 4266b7f9-08bc-469d-b4e8-a88b751a164e. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.439367] env[62183]: DEBUG nova.network.neutron [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Updating instance_info_cache with network_info: [{"id": "4266b7f9-08bc-469d-b4e8-a88b751a164e", "address": "fa:16:3e:13:48:39", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4266b7f9-08", "ovs_interfaceid": "4266b7f9-08bc-469d-b4e8-a88b751a164e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.463703] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387100, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.539124] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.539765] env[62183]: DEBUG nova.compute.manager [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.543605] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.181s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.609412] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387103, 'name': CreateVM_Task, 'duration_secs': 0.377484} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.609412] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.609412] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.609412] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.609412] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.609412] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92a1d09f-6d4f-46a9-8b90-4b118f7b8f23 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.615024] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 871.615024] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52b57ef0-2a02-cc11-955d-09353fd5e8ec" [ 871.615024] env[62183]: _type = "Task" [ 871.615024] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.627022] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b57ef0-2a02-cc11-955d-09353fd5e8ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.944919] env[62183]: DEBUG oslo_concurrency.lockutils [req-f62bab0d-cc4c-4f05-9347-14217cffb034 req-51b124f7-70f4-42ab-bdec-cf76361f6109 service nova] Releasing lock "refresh_cache-843d35ff-6871-4e2a-8e03-d7229a0d8246" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.961339] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387100, 'name': ReconfigVM_Task, 'duration_secs': 0.871285} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.961651] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 1b0593f2-b712-4926-bfb8-30c92070ff6a/1b0593f2-b712-4926-bfb8-30c92070ff6a.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.963373] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-591dec39-fbd7-4405-bd2a-8b989bc86ba0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.969835] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 871.969835] env[62183]: value = "task-1387104" [ 871.969835] env[62183]: _type = "Task" [ 871.969835] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.982740] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387104, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.048487] env[62183]: DEBUG nova.compute.utils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.055263] env[62183]: INFO nova.compute.claims [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.063290] env[62183]: DEBUG nova.compute.manager [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.063417] env[62183]: DEBUG nova.network.neutron [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 872.112943] env[62183]: DEBUG nova.policy [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7deeb5ccaca42e29f0cb145a17e1097', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37d4936e2dc34e9d9ed73455081ee4ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.125237] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b57ef0-2a02-cc11-955d-09353fd5e8ec, 'name': SearchDatastore_Task, 'duration_secs': 0.024767} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.125595] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.125854] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.126139] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.126339] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.126484] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.126775] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06cba478-4edd-4b6d-84a8-cdcda85d83c4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.138485] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.138810] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 872.139638] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d56a0532-24be-449c-bcf9-7a699627b6cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.145089] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 872.145089] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]523b5345-a6ef-4e97-dbba-4b2e779915e8" [ 872.145089] env[62183]: _type = "Task" [ 872.145089] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.154189] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]523b5345-a6ef-4e97-dbba-4b2e779915e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.278875] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.278875] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.484946] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387104, 'name': Rename_Task, 'duration_secs': 0.21148} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.484946] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.484946] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-213b116c-9a6d-406a-b8b7-355a64076c4d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.490152] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 872.490152] env[62183]: value = "task-1387105" [ 872.490152] env[62183]: _type = "Task" [ 872.490152] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.495805] env[62183]: DEBUG nova.network.neutron [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Successfully created port: 1d3eaacd-3e0b-4985-baa2-11f90dba863c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.506094] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387105, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.567888] env[62183]: DEBUG nova.compute.manager [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.580254] env[62183]: INFO nova.compute.resource_tracker [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating resource usage from migration 777f7ae0-1dd8-44e3-b8ed-0e0a79302012 [ 872.666820] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]523b5345-a6ef-4e97-dbba-4b2e779915e8, 'name': SearchDatastore_Task, 'duration_secs': 0.030535} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.667725] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2cc5537-3eca-4966-a0bc-36f9b5be2ad6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.677437] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 872.677437] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52b9069f-2d6b-c415-d23e-34826f57ef72" [ 872.677437] env[62183]: _type = "Task" [ 872.677437] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.688284] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b9069f-2d6b-c415-d23e-34826f57ef72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.790693] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.791142] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 872.791142] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Rebuilding the list of instances to heal {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 872.852118] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faca6ff1-7583-4fb4-ab65-1798b47a33a9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.860037] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e44a20-978a-4d7e-abcf-1715b028f1aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.893513] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f41927c-8823-4bfe-8713-ba57b3e405bf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.900936] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cde160-2332-4a51-8dbf-e83c16a4c025 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.916945] env[62183]: DEBUG nova.compute.provider_tree [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.002817] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387105, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.048272] env[62183]: DEBUG nova.compute.manager [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Received event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.048500] env[62183]: DEBUG nova.compute.manager [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing instance network info cache due to event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.048684] env[62183]: DEBUG oslo_concurrency.lockutils [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] Acquiring lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.048820] env[62183]: DEBUG oslo_concurrency.lockutils [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] Acquired lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.048982] env[62183]: DEBUG nova.network.neutron [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing network info cache for port dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.191336] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b9069f-2d6b-c415-d23e-34826f57ef72, 'name': SearchDatastore_Task, 'duration_secs': 0.01683} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.191336] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.191336] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 843d35ff-6871-4e2a-8e03-d7229a0d8246/843d35ff-6871-4e2a-8e03-d7229a0d8246.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.191336] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-417d5f75-fe9f-40aa-bca1-ebde77ad2269 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.197989] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 873.197989] env[62183]: value = "task-1387106" [ 873.197989] env[62183]: _type = "Task" [ 873.197989] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.206241] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387106, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.298238] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 873.298390] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 873.298516] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Skipping network cache update for instance because it is Building. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 873.300660] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.300878] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.352242] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.424090] env[62183]: DEBUG nova.scheduler.client.report [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 873.504268] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387105, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.592813] env[62183]: DEBUG nova.compute.manager [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.634573] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='ab1b36fe67d0cc4f3e4ba35f49b20b15',container_format='bare',created_at=2024-10-31T11:11:12Z,direct_url=,disk_format='vmdk',id=2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83,min_disk=1,min_ram=0,name='tempest-test-snap-1156695835',owner='37d4936e2dc34e9d9ed73455081ee4ab',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-31T11:11:27Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.635479] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.635479] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.635479] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.635697] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.635697] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.635980] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.636186] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.636418] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.636672] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.636916] env[62183]: DEBUG nova.virt.hardware [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.638006] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82250b4-b322-4336-a6e7-5c13c5b21e1f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.648652] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0f04ee-ca55-49f9-8342-16ec5cf9c840 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.707795] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387106, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.806521] env[62183]: DEBUG nova.compute.manager [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.933028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.387s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.933028] env[62183]: INFO nova.compute.manager [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Migrating [ 873.933028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.933028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.933028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.064s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.934397] env[62183]: INFO nova.compute.claims [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.964812] env[62183]: DEBUG nova.network.neutron [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updated VIF entry in instance network info cache for port dccdd9b2-d8fa-435b-86d8-ec505124174c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.964812] env[62183]: DEBUG nova.network.neutron [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updating instance_info_cache with network_info: [{"id": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "address": "fa:16:3e:f8:f1:0b", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdccdd9b2-d8", "ovs_interfaceid": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.002819] env[62183]: DEBUG oslo_vmware.api [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387105, 'name': PowerOnVM_Task, 'duration_secs': 1.147243} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.005275] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.005575] env[62183]: INFO nova.compute.manager [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Took 9.90 seconds to spawn the instance on the hypervisor. [ 874.005660] env[62183]: DEBUG nova.compute.manager [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.009551] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9c62dc-659c-4d99-8c45-8cc9647cdd0d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.141035] env[62183]: DEBUG nova.compute.manager [req-9c52f87c-f589-41cc-9fe0-994539340a74 req-0c60f30a-b95d-4f8c-b779-1388fe7fa9f0 service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Received event network-vif-plugged-1d3eaacd-3e0b-4985-baa2-11f90dba863c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.141479] env[62183]: DEBUG oslo_concurrency.lockutils [req-9c52f87c-f589-41cc-9fe0-994539340a74 req-0c60f30a-b95d-4f8c-b779-1388fe7fa9f0 service nova] Acquiring lock "be021940-5fff-429a-9b93-4ad238cdff77-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.143180] env[62183]: DEBUG oslo_concurrency.lockutils [req-9c52f87c-f589-41cc-9fe0-994539340a74 req-0c60f30a-b95d-4f8c-b779-1388fe7fa9f0 service nova] Lock "be021940-5fff-429a-9b93-4ad238cdff77-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.143180] env[62183]: DEBUG oslo_concurrency.lockutils [req-9c52f87c-f589-41cc-9fe0-994539340a74 req-0c60f30a-b95d-4f8c-b779-1388fe7fa9f0 service nova] Lock "be021940-5fff-429a-9b93-4ad238cdff77-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.143180] env[62183]: DEBUG nova.compute.manager [req-9c52f87c-f589-41cc-9fe0-994539340a74 req-0c60f30a-b95d-4f8c-b779-1388fe7fa9f0 service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] No waiting events found dispatching network-vif-plugged-1d3eaacd-3e0b-4985-baa2-11f90dba863c {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.143180] env[62183]: WARNING nova.compute.manager [req-9c52f87c-f589-41cc-9fe0-994539340a74 req-0c60f30a-b95d-4f8c-b779-1388fe7fa9f0 service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Received unexpected event network-vif-plugged-1d3eaacd-3e0b-4985-baa2-11f90dba863c for instance with vm_state building and task_state spawning. [ 874.172395] env[62183]: DEBUG nova.network.neutron [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Successfully updated port: 1d3eaacd-3e0b-4985-baa2-11f90dba863c {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.211848] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387106, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586771} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.212620] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 843d35ff-6871-4e2a-8e03-d7229a0d8246/843d35ff-6871-4e2a-8e03-d7229a0d8246.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 874.212843] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.213106] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21e1ebd6-7144-4489-9a8e-49b5b8cfd3d7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.220022] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 874.220022] env[62183]: value = "task-1387107" [ 874.220022] env[62183]: _type = "Task" [ 874.220022] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.227633] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387107, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.328527] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.439397] env[62183]: INFO nova.compute.rpcapi [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 874.439397] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.467568] env[62183]: DEBUG oslo_concurrency.lockutils [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] Releasing lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.468016] env[62183]: DEBUG nova.compute.manager [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Received event network-changed-4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.468114] env[62183]: DEBUG nova.compute.manager [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Refreshing instance network info cache due to event network-changed-4b70b794-46e6-48de-8932-98875906aa5a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 874.468255] env[62183]: DEBUG oslo_concurrency.lockutils [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] Acquiring lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.468539] env[62183]: DEBUG oslo_concurrency.lockutils [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] Acquired lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.468539] env[62183]: DEBUG nova.network.neutron [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Refreshing network info cache for port 4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.469557] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquired lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.469689] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Forcefully refreshing network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 874.470244] env[62183]: DEBUG nova.objects.instance [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lazy-loading 'info_cache' on Instance uuid 7465df88-c68c-49d3-9a91-6fff0d06957b {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.527886] env[62183]: INFO nova.compute.manager [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Took 28.20 seconds to build instance. [ 874.676489] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "refresh_cache-be021940-5fff-429a-9b93-4ad238cdff77" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.676676] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "refresh_cache-be021940-5fff-429a-9b93-4ad238cdff77" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.676799] env[62183]: DEBUG nova.network.neutron [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.710036] env[62183]: DEBUG oslo_concurrency.lockutils [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.710036] env[62183]: DEBUG oslo_concurrency.lockutils [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.710036] env[62183]: DEBUG oslo_concurrency.lockutils [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.710036] env[62183]: DEBUG oslo_concurrency.lockutils [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.710328] env[62183]: DEBUG oslo_concurrency.lockutils [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.710328] env[62183]: INFO nova.compute.manager [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Terminating instance [ 874.711944] env[62183]: DEBUG nova.compute.manager [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.712144] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.713153] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e234b9-7708-42b7-8c44-3730a6837f1e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.722640] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.726103] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ee36eec-c29e-4445-adb2-321194afc7c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.732295] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387107, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066237} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.733405] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.733798] env[62183]: DEBUG oslo_vmware.api [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 874.733798] env[62183]: value = "task-1387108" [ 874.733798] env[62183]: _type = "Task" [ 874.733798] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.734497] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9eeb6b7-4a31-4291-b78d-b19b91b45374 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.744994] env[62183]: DEBUG oslo_vmware.api [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387108, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.763315] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 843d35ff-6871-4e2a-8e03-d7229a0d8246/843d35ff-6871-4e2a-8e03-d7229a0d8246.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.763652] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-962d26d4-8d03-4611-9a8c-47c0984011df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.782165] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 874.782165] env[62183]: value = "task-1387109" [ 874.782165] env[62183]: _type = "Task" [ 874.782165] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.789797] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387109, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.958422] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.958621] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.958798] env[62183]: DEBUG nova.network.neutron [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.030218] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17deee3e-f349-4d37-804e-dd8d2b23dd78 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.798s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.223807] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014bc09b-9342-44e4-86e5-36145a7b74d9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.232955] env[62183]: DEBUG nova.network.neutron [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.237241] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe5ef2b-d9db-4f6f-ab0c-11c8cfe6d14b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.253813] env[62183]: DEBUG oslo_vmware.api [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387108, 'name': PowerOffVM_Task, 'duration_secs': 0.278887} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.279915] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 875.280157] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.284122] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22a29958-de41-4ea0-ae15-e6f5d8acd73c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.286195] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f3e8053-664b-4178-a632-0d0690bffe77 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.290266] env[62183]: DEBUG nova.compute.manager [req-1b34be37-2ba6-41af-9de4-5e8a349ee70c req-0753f943-99d5-4b32-809e-617931576c27 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Received event network-changed-4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.290266] env[62183]: DEBUG nova.compute.manager [req-1b34be37-2ba6-41af-9de4-5e8a349ee70c req-0753f943-99d5-4b32-809e-617931576c27 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Refreshing instance network info cache due to event network-changed-4b70b794-46e6-48de-8932-98875906aa5a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.290266] env[62183]: DEBUG oslo_concurrency.lockutils [req-1b34be37-2ba6-41af-9de4-5e8a349ee70c req-0753f943-99d5-4b32-809e-617931576c27 service nova] Acquiring lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.303078] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bded1bd-cd03-4df1-9ef6-211a3fbec5f9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.308368] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387109, 'name': ReconfigVM_Task, 'duration_secs': 0.282566} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.309051] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 843d35ff-6871-4e2a-8e03-d7229a0d8246/843d35ff-6871-4e2a-8e03-d7229a0d8246.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.312131] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-564e93d4-9836-4472-8ea2-953f39e713f8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.322841] env[62183]: DEBUG nova.compute.provider_tree [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.329378] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 875.329378] env[62183]: value = "task-1387111" [ 875.329378] env[62183]: _type = "Task" [ 875.329378] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.338516] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387111, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.367450] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.367450] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.367450] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Deleting the datastore file [datastore1] 7b5f50f8-5527-4f29-9fa6-6a598b77ef80 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.367768] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-101d90ea-f2e0-406c-be2e-cfeb813f4bfa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.377180] env[62183]: DEBUG oslo_vmware.api [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 875.377180] env[62183]: value = "task-1387112" [ 875.377180] env[62183]: _type = "Task" [ 875.377180] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.388732] env[62183]: DEBUG oslo_vmware.api [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.540407] env[62183]: DEBUG nova.network.neutron [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Updating instance_info_cache with network_info: [{"id": "1d3eaacd-3e0b-4985-baa2-11f90dba863c", "address": "fa:16:3e:0a:dc:20", "network": {"id": "bef4069c-60bb-4ca3-b068-b7d33e5c9d90", "bridge": "br-int", "label": "tempest-ImagesTestJSON-798269247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37d4936e2dc34e9d9ed73455081ee4ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3eaacd-3e", "ovs_interfaceid": "1d3eaacd-3e0b-4985-baa2-11f90dba863c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.615484] env[62183]: DEBUG nova.network.neutron [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Updated VIF entry in instance network info cache for port 4b70b794-46e6-48de-8932-98875906aa5a. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 875.615795] env[62183]: DEBUG nova.network.neutron [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Updating instance_info_cache with network_info: [{"id": "4b70b794-46e6-48de-8932-98875906aa5a", "address": "fa:16:3e:3a:a3:a7", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b70b794-46", "ovs_interfaceid": "4b70b794-46e6-48de-8932-98875906aa5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.829849] env[62183]: DEBUG nova.scheduler.client.report [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 875.841085] env[62183]: DEBUG nova.network.neutron [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance_info_cache with network_info: [{"id": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "address": "fa:16:3e:2f:6f:39", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2f9a654-1f", "ovs_interfaceid": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.846966] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387111, 'name': Rename_Task, 'duration_secs': 0.144169} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.847240] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.847484] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f2929b2-69be-4152-81fa-15f7a62b820d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.853590] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 875.853590] env[62183]: value = "task-1387113" [ 875.853590] env[62183]: _type = "Task" [ 875.853590] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.864720] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387113, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.890353] env[62183]: DEBUG oslo_vmware.api [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387112, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.455084} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.890682] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.890893] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.891091] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.891271] env[62183]: INFO nova.compute.manager [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Took 1.18 seconds to destroy the instance on the hypervisor. [ 875.891520] env[62183]: DEBUG oslo.service.loopingcall [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.891714] env[62183]: DEBUG nova.compute.manager [-] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.891808] env[62183]: DEBUG nova.network.neutron [-] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 876.043515] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "refresh_cache-be021940-5fff-429a-9b93-4ad238cdff77" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.043850] env[62183]: DEBUG nova.compute.manager [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Instance network_info: |[{"id": "1d3eaacd-3e0b-4985-baa2-11f90dba863c", "address": "fa:16:3e:0a:dc:20", "network": {"id": "bef4069c-60bb-4ca3-b068-b7d33e5c9d90", "bridge": "br-int", "label": "tempest-ImagesTestJSON-798269247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37d4936e2dc34e9d9ed73455081ee4ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3eaacd-3e", "ovs_interfaceid": "1d3eaacd-3e0b-4985-baa2-11f90dba863c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 876.044456] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:dc:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11032cc2-b275-48d2-9c40-9455ea7d49e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d3eaacd-3e0b-4985-baa2-11f90dba863c', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.056613] env[62183]: DEBUG oslo.service.loopingcall [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.056613] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.056747] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2955ded-be5c-49a6-bf3b-733dac84955a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.087578] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.087578] env[62183]: value = "task-1387114" [ 876.087578] env[62183]: _type = "Task" [ 876.087578] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.095389] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387114, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.118511] env[62183]: DEBUG oslo_concurrency.lockutils [req-15d12e66-aefe-4abf-b25b-79ec6e94b3cc req-5f506fdd-8721-4ae9-9280-e24ef108abe7 service nova] Releasing lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.118940] env[62183]: DEBUG oslo_concurrency.lockutils [req-1b34be37-2ba6-41af-9de4-5e8a349ee70c req-0753f943-99d5-4b32-809e-617931576c27 service nova] Acquired lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.119228] env[62183]: DEBUG nova.network.neutron [req-1b34be37-2ba6-41af-9de4-5e8a349ee70c req-0753f943-99d5-4b32-809e-617931576c27 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Refreshing network info cache for port 4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.306331] env[62183]: DEBUG nova.compute.manager [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Received event network-changed-1d3eaacd-3e0b-4985-baa2-11f90dba863c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.306574] env[62183]: DEBUG nova.compute.manager [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Refreshing instance network info cache due to event network-changed-1d3eaacd-3e0b-4985-baa2-11f90dba863c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 876.306796] env[62183]: DEBUG oslo_concurrency.lockutils [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] Acquiring lock "refresh_cache-be021940-5fff-429a-9b93-4ad238cdff77" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.306964] env[62183]: DEBUG oslo_concurrency.lockutils [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] Acquired lock "refresh_cache-be021940-5fff-429a-9b93-4ad238cdff77" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.307143] env[62183]: DEBUG nova.network.neutron [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Refreshing network info cache for port 1d3eaacd-3e0b-4985-baa2-11f90dba863c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.343429] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.344151] env[62183]: DEBUG nova.compute.manager [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 876.350207] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.351686] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.023s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.357204] env[62183]: INFO nova.compute.claims [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.374500] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387113, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.464889] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updating instance_info_cache with network_info: [{"id": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "address": "fa:16:3e:f8:f1:0b", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdccdd9b2-d8", "ovs_interfaceid": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.598497] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387114, 'name': CreateVM_Task, 'duration_secs': 0.34941} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.599642] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.602484] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.602484] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.602484] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.602484] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d51c1fd2-9e3e-4889-98d8-106bbcfbf734 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.609131] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 876.609131] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5255ccd5-8941-0793-081d-bd0f3ae5244d" [ 876.609131] env[62183]: _type = "Task" [ 876.609131] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.618090] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5255ccd5-8941-0793-081d-bd0f3ae5244d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.643227] env[62183]: INFO nova.network.neutron [req-1b34be37-2ba6-41af-9de4-5e8a349ee70c req-0753f943-99d5-4b32-809e-617931576c27 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Port 4b70b794-46e6-48de-8932-98875906aa5a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 876.643457] env[62183]: DEBUG nova.network.neutron [req-1b34be37-2ba6-41af-9de4-5e8a349ee70c req-0753f943-99d5-4b32-809e-617931576c27 service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.850529] env[62183]: DEBUG nova.compute.utils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.851946] env[62183]: DEBUG nova.compute.manager [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.852213] env[62183]: DEBUG nova.network.neutron [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 876.875369] env[62183]: DEBUG oslo_vmware.api [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387113, 'name': PowerOnVM_Task, 'duration_secs': 0.560408} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.875645] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 876.875845] env[62183]: INFO nova.compute.manager [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Took 8.53 seconds to spawn the instance on the hypervisor. [ 876.876101] env[62183]: DEBUG nova.compute.manager [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.880197] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bae6023-cd8f-4eba-9b31-37f5fe46164e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.922799] env[62183]: DEBUG nova.network.neutron [-] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.951000] env[62183]: DEBUG nova.policy [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ae1837d5db145278417f7cdd55a3fea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06e81bd1a81d4009ae2a75fe819f9b7c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.967559] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Releasing lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.967746] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updated the network info_cache for instance {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 876.967970] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.968189] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.968416] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.968567] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.969102] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.969324] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.969496] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 876.969646] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 877.122033] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.123227] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Processing image 2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.123590] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.123766] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "[datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.124121] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.124742] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6da559a7-c6fa-4f4b-bc74-dea5894b06eb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.137855] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.138217] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.139098] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d111a85-b69a-488b-88cd-2873da652cb3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.143720] env[62183]: DEBUG nova.network.neutron [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Updated VIF entry in instance network info cache for port 1d3eaacd-3e0b-4985-baa2-11f90dba863c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 877.143720] env[62183]: DEBUG nova.network.neutron [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Updating instance_info_cache with network_info: [{"id": "1d3eaacd-3e0b-4985-baa2-11f90dba863c", "address": "fa:16:3e:0a:dc:20", "network": {"id": "bef4069c-60bb-4ca3-b068-b7d33e5c9d90", "bridge": "br-int", "label": "tempest-ImagesTestJSON-798269247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37d4936e2dc34e9d9ed73455081ee4ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3eaacd-3e", "ovs_interfaceid": "1d3eaacd-3e0b-4985-baa2-11f90dba863c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.145725] env[62183]: DEBUG oslo_concurrency.lockutils [req-1b34be37-2ba6-41af-9de4-5e8a349ee70c req-0753f943-99d5-4b32-809e-617931576c27 service nova] Releasing lock "refresh_cache-7b5f50f8-5527-4f29-9fa6-6a598b77ef80" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.147506] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 877.147506] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f688f0-a7eb-6f86-37f3-985c4838af3e" [ 877.147506] env[62183]: _type = "Task" [ 877.147506] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.155422] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f688f0-a7eb-6f86-37f3-985c4838af3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.308325] env[62183]: DEBUG nova.compute.manager [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.308636] env[62183]: DEBUG nova.compute.manager [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing instance network info cache due to event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.308995] env[62183]: DEBUG oslo_concurrency.lockutils [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] Acquiring lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.309271] env[62183]: DEBUG oslo_concurrency.lockutils [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] Acquired lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.309638] env[62183]: DEBUG nova.network.neutron [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.324474] env[62183]: DEBUG nova.network.neutron [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Successfully created port: 8d0bfbea-f6ad-41ee-a0f1-4ccad145901f {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.354985] env[62183]: DEBUG nova.compute.manager [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 877.401836] env[62183]: INFO nova.compute.manager [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Took 27.13 seconds to build instance. [ 877.430132] env[62183]: INFO nova.compute.manager [-] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Took 1.54 seconds to deallocate network for instance. [ 877.472401] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.613477] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146032a0-2944-4f99-996a-fa555cd7d54e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.624228] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09603fc6-df94-40e8-9a63-f489697e06c4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.661075] env[62183]: DEBUG oslo_concurrency.lockutils [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] Releasing lock "refresh_cache-be021940-5fff-429a-9b93-4ad238cdff77" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.661075] env[62183]: DEBUG nova.compute.manager [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Received event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.661075] env[62183]: DEBUG nova.compute.manager [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing instance network info cache due to event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.661075] env[62183]: DEBUG oslo_concurrency.lockutils [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] Acquiring lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.661075] env[62183]: DEBUG oslo_concurrency.lockutils [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] Acquired lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.661309] env[62183]: DEBUG nova.network.neutron [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing network info cache for port dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.666059] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c2771c-575d-49cb-b783-eb7941aeff61 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.679019] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Preparing fetch location {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 877.679019] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Fetch image to [datastore2] OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836/OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836.vmdk {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 877.679019] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Downloading stream optimized image 2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83 to [datastore2] OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836/OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836.vmdk on the data store datastore2 as vApp {{(pid=62183) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 877.679019] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Downloading image file data 2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83 to the ESX as VM named 'OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836' {{(pid=62183) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 877.684596] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3a0600-b745-4889-97ea-4eff3067744c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.699210] env[62183]: DEBUG nova.compute.provider_tree [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.777650] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 877.777650] env[62183]: value = "resgroup-9" [ 877.777650] env[62183]: _type = "ResourcePool" [ 877.777650] env[62183]: }. {{(pid=62183) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 877.778433] env[62183]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b6ab255b-1307-4638-8b68-60e36f6bf32b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.799849] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lease: (returnval){ [ 877.799849] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52cce7f5-9947-7b6b-04c6-aff75560282b" [ 877.799849] env[62183]: _type = "HttpNfcLease" [ 877.799849] env[62183]: } obtained for vApp import into resource pool (val){ [ 877.799849] env[62183]: value = "resgroup-9" [ 877.799849] env[62183]: _type = "ResourcePool" [ 877.799849] env[62183]: }. {{(pid=62183) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 877.800152] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the lease: (returnval){ [ 877.800152] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52cce7f5-9947-7b6b-04c6-aff75560282b" [ 877.800152] env[62183]: _type = "HttpNfcLease" [ 877.800152] env[62183]: } to be ready. {{(pid=62183) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 877.807047] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 877.807047] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52cce7f5-9947-7b6b-04c6-aff75560282b" [ 877.807047] env[62183]: _type = "HttpNfcLease" [ 877.807047] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 877.877069] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea72be69-a1ab-419d-91fe-03e18791d1d5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.894610] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance '503786ca-dba4-43c1-9a25-9f1cbac9a6a4' progress to 0 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 877.902030] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9883952d-0ed5-4522-b7e1-5ec2fadd67aa tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.641s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.940172] env[62183]: DEBUG oslo_concurrency.lockutils [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.064793] env[62183]: DEBUG nova.network.neutron [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updated VIF entry in instance network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.065182] env[62183]: DEBUG nova.network.neutron [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.202175] env[62183]: DEBUG nova.scheduler.client.report [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 878.308016] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 878.308016] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52cce7f5-9947-7b6b-04c6-aff75560282b" [ 878.308016] env[62183]: _type = "HttpNfcLease" [ 878.308016] env[62183]: } is ready. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 878.308343] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 878.308343] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52cce7f5-9947-7b6b-04c6-aff75560282b" [ 878.308343] env[62183]: _type = "HttpNfcLease" [ 878.308343] env[62183]: }. {{(pid=62183) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 878.309119] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8848d49e-7462-4e3c-9401-9b5625f6b6ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.316943] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524578dd-7cd7-b9cf-96ec-f8e7eea7b769/disk-0.vmdk from lease info. {{(pid=62183) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 878.317057] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524578dd-7cd7-b9cf-96ec-f8e7eea7b769/disk-0.vmdk. {{(pid=62183) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 878.377623] env[62183]: DEBUG nova.compute.manager [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 878.385526] env[62183]: DEBUG nova.compute.manager [req-6ae709de-5373-4658-839f-ddcb94581fe2 req-6a2deca8-8b83-4464-ba59-c6eeb9a35c3d service nova] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Received event network-vif-deleted-4b70b794-46e6-48de-8932-98875906aa5a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.385767] env[62183]: DEBUG nova.compute.manager [req-6ae709de-5373-4658-839f-ddcb94581fe2 req-6a2deca8-8b83-4464-ba59-c6eeb9a35c3d service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Received event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.386082] env[62183]: DEBUG nova.compute.manager [req-6ae709de-5373-4658-839f-ddcb94581fe2 req-6a2deca8-8b83-4464-ba59-c6eeb9a35c3d service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing instance network info cache due to event network-changed-dccdd9b2-d8fa-435b-86d8-ec505124174c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.386292] env[62183]: DEBUG oslo_concurrency.lockutils [req-6ae709de-5373-4658-839f-ddcb94581fe2 req-6a2deca8-8b83-4464-ba59-c6eeb9a35c3d service nova] Acquiring lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.390586] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cc4a6e08-a325-41f4-acaa-8b3c622a4136 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.402827] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 878.405463] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 878.405737] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 878.405997] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.406229] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 878.406403] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.406585] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 878.406865] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 878.407063] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 878.407281] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 878.407480] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 878.407689] env[62183]: DEBUG nova.virt.hardware [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 878.408057] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02f9fdeb-490d-4615-b4af-69f2c436a515 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.410743] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c3c0d8-7eb7-4894-a67c-03addca96950 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.421100] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735313b2-94ea-4616-b1cd-4113f3089ab8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.425922] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 878.425922] env[62183]: value = "task-1387116" [ 878.425922] env[62183]: _type = "Task" [ 878.425922] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.446526] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387116, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.479950] env[62183]: DEBUG nova.network.neutron [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updated VIF entry in instance network info cache for port dccdd9b2-d8fa-435b-86d8-ec505124174c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.480413] env[62183]: DEBUG nova.network.neutron [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updating instance_info_cache with network_info: [{"id": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "address": "fa:16:3e:f8:f1:0b", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdccdd9b2-d8", "ovs_interfaceid": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.563766] env[62183]: DEBUG oslo_concurrency.lockutils [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "843d35ff-6871-4e2a-8e03-d7229a0d8246" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.563766] env[62183]: DEBUG oslo_concurrency.lockutils [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.564069] env[62183]: DEBUG nova.compute.manager [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.567393] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d204c3-544f-41e3-9218-199df94d548d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.572612] env[62183]: DEBUG oslo_concurrency.lockutils [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] Releasing lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.572873] env[62183]: DEBUG nova.compute.manager [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received event network-changed-999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.573056] env[62183]: DEBUG nova.compute.manager [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing instance network info cache due to event network-changed-999bf20b-d0e4-4229-a150-ec7c31e38cc6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.573267] env[62183]: DEBUG oslo_concurrency.lockutils [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] Acquiring lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.573418] env[62183]: DEBUG oslo_concurrency.lockutils [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] Acquired lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.573658] env[62183]: DEBUG nova.network.neutron [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing network info cache for port 999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.581288] env[62183]: DEBUG nova.compute.manager [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62183) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 878.581288] env[62183]: DEBUG nova.objects.instance [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lazy-loading 'flavor' on Instance uuid 843d35ff-6871-4e2a-8e03-d7229a0d8246 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.707846] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.708475] env[62183]: DEBUG nova.compute.manager [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 878.714301] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.240s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.714301] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.714301] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 878.714301] env[62183]: DEBUG oslo_concurrency.lockutils [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.773s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.714301] env[62183]: DEBUG nova.objects.instance [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lazy-loading 'resources' on Instance uuid 7b5f50f8-5527-4f29-9fa6-6a598b77ef80 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.715572] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df064db-7e56-49de-86c2-21fcdba04db2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.729975] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed029af9-a587-4ff7-8f5d-94ce6ee34aa7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.752104] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cda1393-8543-410a-80dd-aa5306ae0b59 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.761850] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-892969c3-8e35-47f5-bc3c-d0b1c6e72e15 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.798194] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181349MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 878.798355] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.939918] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387116, 'name': PowerOffVM_Task, 'duration_secs': 0.35245} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.940268] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 878.940493] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance '503786ca-dba4-43c1-9a25-9f1cbac9a6a4' progress to 17 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 878.983364] env[62183]: DEBUG oslo_concurrency.lockutils [req-6152ce2d-b77c-4308-b608-ffe75d1da42d req-7f5d85c0-0a43-460f-be02-befe6112adc3 service nova] Releasing lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.983906] env[62183]: DEBUG oslo_concurrency.lockutils [req-6ae709de-5373-4658-839f-ddcb94581fe2 req-6a2deca8-8b83-4464-ba59-c6eeb9a35c3d service nova] Acquired lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.984128] env[62183]: DEBUG nova.network.neutron [req-6ae709de-5373-4658-839f-ddcb94581fe2 req-6a2deca8-8b83-4464-ba59-c6eeb9a35c3d service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Refreshing network info cache for port dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.085835] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 879.086200] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b6b47ffc-6276-432f-86a5-1f3b9818d5ff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.093830] env[62183]: DEBUG oslo_vmware.api [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 879.093830] env[62183]: value = "task-1387117" [ 879.093830] env[62183]: _type = "Task" [ 879.093830] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.108157] env[62183]: DEBUG oslo_vmware.api [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387117, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.137195] env[62183]: DEBUG nova.network.neutron [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Successfully updated port: 8d0bfbea-f6ad-41ee-a0f1-4ccad145901f {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.164312] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Completed reading data from the image iterator. {{(pid=62183) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 879.164513] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524578dd-7cd7-b9cf-96ec-f8e7eea7b769/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 879.166470] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0ad613-947e-4d39-abd8-64b9121702a8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.179722] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524578dd-7cd7-b9cf-96ec-f8e7eea7b769/disk-0.vmdk is in state: ready. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 879.179722] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524578dd-7cd7-b9cf-96ec-f8e7eea7b769/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 879.180071] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-bfbe5e03-7d67-4073-8d8c-21e09dbabff3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.216956] env[62183]: DEBUG nova.compute.utils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.222032] env[62183]: DEBUG nova.compute.manager [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.222032] env[62183]: DEBUG nova.network.neutron [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 879.310813] env[62183]: DEBUG nova.policy [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88efcb2057e84413806f9ed809c260e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eda7e297c551407295b3c67fa0f1dbc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 879.405919] env[62183]: DEBUG oslo_vmware.rw_handles [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/524578dd-7cd7-b9cf-96ec-f8e7eea7b769/disk-0.vmdk. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 879.406289] env[62183]: INFO nova.virt.vmwareapi.images [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Downloaded image file data 2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83 [ 879.407935] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46772aa-9548-451c-b23b-c6f86dfac3c8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.435689] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d886c1c-e798-4c64-b1f9-bf7681ca907f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.448819] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.448819] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.448819] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.448819] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.450204] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.454022] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.454022] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.454022] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.454022] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.454022] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.454606] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.461831] env[62183]: DEBUG nova.network.neutron [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updated VIF entry in instance network info cache for port 999bf20b-d0e4-4229-a150-ec7c31e38cc6. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.462344] env[62183]: DEBUG nova.network.neutron [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updating instance_info_cache with network_info: [{"id": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "address": "fa:16:3e:46:24:80", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap999bf20b-d0", "ovs_interfaceid": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.464475] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef2be7f0-ba3e-4dea-9d78-03dbdc0b9f3e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.476463] env[62183]: DEBUG nova.compute.manager [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received event network-changed-999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.476855] env[62183]: DEBUG nova.compute.manager [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing instance network info cache due to event network-changed-999bf20b-d0e4-4229-a150-ec7c31e38cc6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 879.477183] env[62183]: DEBUG oslo_concurrency.lockutils [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] Acquiring lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.478974] env[62183]: DEBUG oslo_concurrency.lockutils [req-eb892383-ebf1-4b4e-a8a0-12542ddfc89c req-c1a9e65a-c59a-4962-84a5-bddc7bc7c262 service nova] Releasing lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.479887] env[62183]: DEBUG oslo_concurrency.lockutils [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] Acquired lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.480307] env[62183]: DEBUG nova.network.neutron [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing network info cache for port 999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.490216] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 879.490216] env[62183]: value = "task-1387119" [ 879.490216] env[62183]: _type = "Task" [ 879.490216] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.500553] env[62183]: INFO nova.virt.vmwareapi.images [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] The imported VM was unregistered [ 879.503348] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Caching image {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 879.503744] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Creating directory with path [datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83 {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.508097] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-690b5894-1e12-44d6-8204-28a2f0aafcdb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.510271] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387119, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.512074] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59fe325-764e-4092-bbb8-942764853281 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.519551] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559c9384-c108-4ae6-9853-70526d5e01d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.523842] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Created directory with path [datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83 {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.524205] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836/OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836.vmdk to [datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83.vmdk. {{(pid=62183) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 879.524871] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-dbd6bf49-39e7-49bd-9157-3e8f0567e812 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.550135] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "interface-ac555ffc-ce4e-4650-97fd-c26a3246fe4b-e7836918-44a9-4db8-8daf-35c2032e5cdb" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.550429] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "interface-ac555ffc-ce4e-4650-97fd-c26a3246fe4b-e7836918-44a9-4db8-8daf-35c2032e5cdb" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.550795] env[62183]: DEBUG nova.objects.instance [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lazy-loading 'flavor' on Instance uuid ac555ffc-ce4e-4650-97fd-c26a3246fe4b {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.556101] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9de1deb-633f-4feb-9d0e-4cf571083bc2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.560471] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 879.560471] env[62183]: value = "task-1387120" [ 879.560471] env[62183]: _type = "Task" [ 879.560471] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.568983] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec295330-892b-44e0-92e2-b4dc08e24b87 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.576560] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387120, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.589108] env[62183]: DEBUG nova.compute.provider_tree [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.603979] env[62183]: DEBUG oslo_vmware.api [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387117, 'name': PowerOffVM_Task, 'duration_secs': 0.329236} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.604321] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.604544] env[62183]: DEBUG nova.compute.manager [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.605382] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7284e9b0-2a46-4e2d-b4cb-8927708f461a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.641218] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "refresh_cache-ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.641388] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "refresh_cache-ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.641547] env[62183]: DEBUG nova.network.neutron [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.728475] env[62183]: DEBUG nova.compute.manager [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 879.775850] env[62183]: DEBUG nova.network.neutron [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Successfully created port: 4e6d3d8d-6821-471a-bca5-29acccb74622 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 879.869924] env[62183]: DEBUG nova.network.neutron [req-6ae709de-5373-4658-839f-ddcb94581fe2 req-6a2deca8-8b83-4464-ba59-c6eeb9a35c3d service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updated VIF entry in instance network info cache for port dccdd9b2-d8fa-435b-86d8-ec505124174c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.870352] env[62183]: DEBUG nova.network.neutron [req-6ae709de-5373-4658-839f-ddcb94581fe2 req-6a2deca8-8b83-4464-ba59-c6eeb9a35c3d service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updating instance_info_cache with network_info: [{"id": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "address": "fa:16:3e:f8:f1:0b", "network": {"id": "61fad9ec-2983-4242-9892-d1899830c4fb", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-768141839-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c9dfd905ec14f2bbbe2b98692222f17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdccdd9b2-d8", "ovs_interfaceid": "dccdd9b2-d8fa-435b-86d8-ec505124174c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.004126] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387119, 'name': ReconfigVM_Task, 'duration_secs': 0.25317} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.006771] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance '503786ca-dba4-43c1-9a25-9f1cbac9a6a4' progress to 33 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 880.075203] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387120, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.092647] env[62183]: DEBUG nova.scheduler.client.report [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 880.118824] env[62183]: DEBUG oslo_concurrency.lockutils [None req-145bafb5-61b1-415d-b34d-8bbb60869729 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.197382] env[62183]: DEBUG nova.network.neutron [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.243965] env[62183]: DEBUG nova.objects.instance [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lazy-loading 'pci_requests' on Instance uuid ac555ffc-ce4e-4650-97fd-c26a3246fe4b {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.295053] env[62183]: DEBUG nova.network.neutron [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updated VIF entry in instance network info cache for port 999bf20b-d0e4-4229-a150-ec7c31e38cc6. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.295509] env[62183]: DEBUG nova.network.neutron [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updating instance_info_cache with network_info: [{"id": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "address": "fa:16:3e:46:24:80", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap999bf20b-d0", "ovs_interfaceid": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.373423] env[62183]: DEBUG oslo_concurrency.lockutils [req-6ae709de-5373-4658-839f-ddcb94581fe2 req-6a2deca8-8b83-4464-ba59-c6eeb9a35c3d service nova] Releasing lock "refresh_cache-7465df88-c68c-49d3-9a91-6fff0d06957b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.451938] env[62183]: DEBUG nova.network.neutron [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Updating instance_info_cache with network_info: [{"id": "8d0bfbea-f6ad-41ee-a0f1-4ccad145901f", "address": "fa:16:3e:07:ab:a2", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d0bfbea-f6", "ovs_interfaceid": "8d0bfbea-f6ad-41ee-a0f1-4ccad145901f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.490597] env[62183]: DEBUG nova.compute.manager [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Received event network-vif-plugged-8d0bfbea-f6ad-41ee-a0f1-4ccad145901f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.491879] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] Acquiring lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.491879] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.491879] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.491879] env[62183]: DEBUG nova.compute.manager [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] No waiting events found dispatching network-vif-plugged-8d0bfbea-f6ad-41ee-a0f1-4ccad145901f {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.491879] env[62183]: WARNING nova.compute.manager [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Received unexpected event network-vif-plugged-8d0bfbea-f6ad-41ee-a0f1-4ccad145901f for instance with vm_state building and task_state spawning. [ 880.492201] env[62183]: DEBUG nova.compute.manager [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Received event network-changed-8d0bfbea-f6ad-41ee-a0f1-4ccad145901f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.492201] env[62183]: DEBUG nova.compute.manager [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Refreshing instance network info cache due to event network-changed-8d0bfbea-f6ad-41ee-a0f1-4ccad145901f. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.492201] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] Acquiring lock "refresh_cache-ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.514864] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.515111] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.515278] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.515460] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.515618] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.515780] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.515985] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.516185] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.516330] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.516494] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.516669] env[62183]: DEBUG nova.virt.hardware [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.522153] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Reconfiguring VM instance instance-00000049 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 880.522475] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c14eae67-662f-454e-8eb4-bd81fe26f95a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.541980] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 880.541980] env[62183]: value = "task-1387121" [ 880.541980] env[62183]: _type = "Task" [ 880.541980] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.550181] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387121, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.573805] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387120, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.600963] env[62183]: DEBUG oslo_concurrency.lockutils [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.888s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.604399] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.805s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.626513] env[62183]: INFO nova.scheduler.client.report [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Deleted allocations for instance 7b5f50f8-5527-4f29-9fa6-6a598b77ef80 [ 880.743891] env[62183]: DEBUG nova.compute.manager [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.750145] env[62183]: DEBUG nova.objects.base [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 880.750408] env[62183]: DEBUG nova.network.neutron [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.770502] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.770990] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.770990] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.771176] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.771327] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.771519] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.771731] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.771892] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.772077] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.772275] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.772510] env[62183]: DEBUG nova.virt.hardware [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.773451] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e31bbc-4361-444d-aa0e-0c889aa0a0bd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.782340] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1a65a1-55a2-4b23-b867-aa0f9f98237f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.798905] env[62183]: DEBUG oslo_concurrency.lockutils [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] Releasing lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.799129] env[62183]: DEBUG nova.compute.manager [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.799308] env[62183]: DEBUG nova.compute.manager [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing instance network info cache due to event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.799528] env[62183]: DEBUG oslo_concurrency.lockutils [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] Acquiring lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.799674] env[62183]: DEBUG oslo_concurrency.lockutils [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] Acquired lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.799830] env[62183]: DEBUG nova.network.neutron [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.849681] env[62183]: DEBUG nova.policy [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e0eee67c7954c80b4e1ea43b0abbcb5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebcc716f958942b588a6bfde78d2c00d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.954369] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "refresh_cache-ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.954745] env[62183]: DEBUG nova.compute.manager [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Instance network_info: |[{"id": "8d0bfbea-f6ad-41ee-a0f1-4ccad145901f", "address": "fa:16:3e:07:ab:a2", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d0bfbea-f6", "ovs_interfaceid": "8d0bfbea-f6ad-41ee-a0f1-4ccad145901f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.955092] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] Acquired lock "refresh_cache-ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.955328] env[62183]: DEBUG nova.network.neutron [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Refreshing network info cache for port 8d0bfbea-f6ad-41ee-a0f1-4ccad145901f {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.956675] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:ab:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '205fb402-8eaf-4b61-8f57-8f216024179a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8d0bfbea-f6ad-41ee-a0f1-4ccad145901f', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.965363] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Creating folder: Project (06e81bd1a81d4009ae2a75fe819f9b7c). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.966531] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a639f711-99d0-42d5-b59c-9327fdb1b87d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.977949] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Created folder: Project (06e81bd1a81d4009ae2a75fe819f9b7c) in parent group-v294392. [ 880.978197] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Creating folder: Instances. Parent ref: group-v294464. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.978452] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de2dfad8-e9eb-4225-8b6d-89f98967e6ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.988348] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Created folder: Instances in parent group-v294464. [ 880.988614] env[62183]: DEBUG oslo.service.loopingcall [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.988907] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.989059] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96679f25-358c-451a-b7dd-3ba94d25d682 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.009562] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.009562] env[62183]: value = "task-1387124" [ 881.009562] env[62183]: _type = "Task" [ 881.009562] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.018649] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387124, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.051902] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387121, 'name': ReconfigVM_Task, 'duration_secs': 0.29985} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.051902] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Reconfigured VM instance instance-00000049 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 881.052953] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec72825-ae1f-4a36-98aa-adea7f355d59 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.077174] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 503786ca-dba4-43c1-9a25-9f1cbac9a6a4/503786ca-dba4-43c1-9a25-9f1cbac9a6a4.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.081017] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ffd56dc-9d47-4f0c-8606-5bf35c19be3a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.099869] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387120, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.101878] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 881.101878] env[62183]: value = "task-1387125" [ 881.101878] env[62183]: _type = "Task" [ 881.101878] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.120264] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387125, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.134934] env[62183]: DEBUG oslo_concurrency.lockutils [None req-649c1d3a-0eda-4157-b2ea-becddc78f774 tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7b5f50f8-5527-4f29-9fa6-6a598b77ef80" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.427s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.162963] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "843d35ff-6871-4e2a-8e03-d7229a0d8246" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.163261] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.163757] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "843d35ff-6871-4e2a-8e03-d7229a0d8246-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.163757] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.163757] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.166220] env[62183]: INFO nova.compute.manager [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Terminating instance [ 881.168166] env[62183]: DEBUG nova.compute.manager [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.168374] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.169295] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251a0b10-bd99-4cbf-be68-f73fddb5765f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.179881] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.179881] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e90ac2d-f9c9-485c-b875-9749bda40269 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.261688] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.261688] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.261688] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleting the datastore file [datastore2] 843d35ff-6871-4e2a-8e03-d7229a0d8246 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.261908] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10a7e900-0620-4488-931f-ccac371798a1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.271420] env[62183]: DEBUG oslo_vmware.api [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 881.271420] env[62183]: value = "task-1387127" [ 881.271420] env[62183]: _type = "Task" [ 881.271420] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.283765] env[62183]: DEBUG oslo_vmware.api [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387127, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.524224] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387124, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.581258] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387120, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.615020] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387125, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.623054] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Applying migration context for instance 503786ca-dba4-43c1-9a25-9f1cbac9a6a4 as it has an incoming, in-progress migration 777f7ae0-1dd8-44e3-b8ed-0e0a79302012. Migration status is migrating {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 881.624950] env[62183]: INFO nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating resource usage from migration 777f7ae0-1dd8-44e3-b8ed-0e0a79302012 [ 881.638019] env[62183]: DEBUG nova.compute.manager [req-6e47a0fb-c839-4272-afe0-1f4e7baedca2 req-88491568-d96b-41ef-9ca5-eaa6cc388372 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Received event network-vif-plugged-4e6d3d8d-6821-471a-bca5-29acccb74622 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.638251] env[62183]: DEBUG oslo_concurrency.lockutils [req-6e47a0fb-c839-4272-afe0-1f4e7baedca2 req-88491568-d96b-41ef-9ca5-eaa6cc388372 service nova] Acquiring lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.638469] env[62183]: DEBUG oslo_concurrency.lockutils [req-6e47a0fb-c839-4272-afe0-1f4e7baedca2 req-88491568-d96b-41ef-9ca5-eaa6cc388372 service nova] Lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.638642] env[62183]: DEBUG oslo_concurrency.lockutils [req-6e47a0fb-c839-4272-afe0-1f4e7baedca2 req-88491568-d96b-41ef-9ca5-eaa6cc388372 service nova] Lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.638812] env[62183]: DEBUG nova.compute.manager [req-6e47a0fb-c839-4272-afe0-1f4e7baedca2 req-88491568-d96b-41ef-9ca5-eaa6cc388372 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] No waiting events found dispatching network-vif-plugged-4e6d3d8d-6821-471a-bca5-29acccb74622 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.638978] env[62183]: WARNING nova.compute.manager [req-6e47a0fb-c839-4272-afe0-1f4e7baedca2 req-88491568-d96b-41ef-9ca5-eaa6cc388372 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Received unexpected event network-vif-plugged-4e6d3d8d-6821-471a-bca5-29acccb74622 for instance with vm_state building and task_state spawning. [ 881.646888] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 7465df88-c68c-49d3-9a91-6fff0d06957b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.647058] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance ac555ffc-ce4e-4650-97fd-c26a3246fe4b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.647186] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 850d40d8-565a-49a2-a27f-3de2a8dc7e30 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.647304] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.647426] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance e3145bd7-85b2-4cc7-9d97-3e36a59b89cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.647540] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a948464b-63aa-4bc8-9885-228049e96d37 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.647654] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance fb348784-62a0-4d1f-ac7f-f176f3da0dd9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.647765] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 1b0593f2-b712-4926-bfb8-30c92070ff6a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.647875] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 843d35ff-6871-4e2a-8e03-d7229a0d8246 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.647984] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance be021940-5fff-429a-9b93-4ad238cdff77 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.648113] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Migration 777f7ae0-1dd8-44e3-b8ed-0e0a79302012 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 881.648263] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 503786ca-dba4-43c1-9a25-9f1cbac9a6a4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.648381] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.648546] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 881.649265] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 881.649265] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 881.785395] env[62183]: DEBUG oslo_vmware.api [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387127, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.853122] env[62183]: DEBUG nova.network.neutron [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Updated VIF entry in instance network info cache for port 8d0bfbea-f6ad-41ee-a0f1-4ccad145901f. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.853122] env[62183]: DEBUG nova.network.neutron [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Updating instance_info_cache with network_info: [{"id": "8d0bfbea-f6ad-41ee-a0f1-4ccad145901f", "address": "fa:16:3e:07:ab:a2", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d0bfbea-f6", "ovs_interfaceid": "8d0bfbea-f6ad-41ee-a0f1-4ccad145901f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.862143] env[62183]: DEBUG nova.network.neutron [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updated VIF entry in instance network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.862547] env[62183]: DEBUG nova.network.neutron [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.868333] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5847b1ad-4772-4e92-8acf-4c3c373c4ff8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.879217] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ed6b01-4ab5-4263-aabc-56c60f9a9da0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.912460] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557bbbec-d54f-46a6-9d4a-ece4be3e42de {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.920692] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f0c036-f0e7-4e60-85d0-29a31a628497 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.936610] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.021795] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387124, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.079859] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387120, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.397509} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.079859] env[62183]: INFO nova.virt.vmwareapi.ds_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836/OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836.vmdk to [datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83.vmdk. [ 882.080119] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Cleaning up location [datastore2] OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 882.080248] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_e67fac1c-bbed-4c12-af4e-ef3915cb3836 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.080492] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b0a4fee-25ae-4437-8b7a-01ba3e8e6cde {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.087043] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 882.087043] env[62183]: value = "task-1387128" [ 882.087043] env[62183]: _type = "Task" [ 882.087043] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.094690] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.112840] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387125, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.264697] env[62183]: DEBUG nova.network.neutron [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Successfully updated port: 4e6d3d8d-6821-471a-bca5-29acccb74622 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.284358] env[62183]: DEBUG oslo_vmware.api [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387127, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.72396} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.284358] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.284514] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.284711] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.285061] env[62183]: INFO nova.compute.manager [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Took 1.12 seconds to destroy the instance on the hypervisor. [ 882.285151] env[62183]: DEBUG oslo.service.loopingcall [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.285344] env[62183]: DEBUG nova.compute.manager [-] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 882.285464] env[62183]: DEBUG nova.network.neutron [-] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.358677] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d145708-57ed-4903-8163-6fb5fdcdca94 req-8d6893f4-bb26-4bda-b4d9-abac460cfb0d service nova] Releasing lock "refresh_cache-ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.365529] env[62183]: DEBUG oslo_concurrency.lockutils [req-13b29b61-4b7c-4034-b490-d559cb3a18f6 req-1dc9e350-c852-4d0a-beac-4bf48860c0f6 service nova] Releasing lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.439861] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 882.525018] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387124, 'name': CreateVM_Task, 'duration_secs': 1.154089} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.525602] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.526557] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.526809] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.527277] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.527634] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-027d10cd-6887-4e60-a32f-01909f2b1859 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.533390] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 882.533390] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f84b48-c6bc-65a2-b5cc-d3fd96434dbd" [ 882.533390] env[62183]: _type = "Task" [ 882.533390] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.541680] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f84b48-c6bc-65a2-b5cc-d3fd96434dbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.599504] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387128, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165839} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.599791] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.599978] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "[datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.600245] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83.vmdk to [datastore2] be021940-5fff-429a-9b93-4ad238cdff77/be021940-5fff-429a-9b93-4ad238cdff77.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.600524] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc670fc6-97c6-42a3-aeba-e2ef9d4622f7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.609743] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 882.609743] env[62183]: value = "task-1387129" [ 882.609743] env[62183]: _type = "Task" [ 882.609743] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.617626] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387125, 'name': ReconfigVM_Task, 'duration_secs': 1.083224} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.618197] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 503786ca-dba4-43c1-9a25-9f1cbac9a6a4/503786ca-dba4-43c1-9a25-9f1cbac9a6a4.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.618489] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance '503786ca-dba4-43c1-9a25-9f1cbac9a6a4' progress to 50 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 882.624407] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.680034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "7465df88-c68c-49d3-9a91-6fff0d06957b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.680034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7465df88-c68c-49d3-9a91-6fff0d06957b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.680034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "7465df88-c68c-49d3-9a91-6fff0d06957b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.680034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7465df88-c68c-49d3-9a91-6fff0d06957b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.680699] env[62183]: DEBUG oslo_concurrency.lockutils [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7465df88-c68c-49d3-9a91-6fff0d06957b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.680750] env[62183]: INFO nova.compute.manager [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Terminating instance [ 882.682891] env[62183]: DEBUG nova.compute.manager [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.683093] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.683907] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4390d06-1c08-494a-859d-7a9e10be31df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.695148] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.695148] env[62183]: DEBUG nova.network.neutron [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Successfully updated port: e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.695148] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5917423-a841-4b86-8941-7eee55a6e205 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.700703] env[62183]: DEBUG oslo_vmware.api [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 882.700703] env[62183]: value = "task-1387130" [ 882.700703] env[62183]: _type = "Task" [ 882.700703] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.712059] env[62183]: DEBUG oslo_vmware.api [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387130, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.770931] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "refresh_cache-71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.770931] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "refresh_cache-71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.770931] env[62183]: DEBUG nova.network.neutron [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.792502] env[62183]: DEBUG nova.compute.manager [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Received event network-changed-4e6d3d8d-6821-471a-bca5-29acccb74622 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.792711] env[62183]: DEBUG nova.compute.manager [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Refreshing instance network info cache due to event network-changed-4e6d3d8d-6821-471a-bca5-29acccb74622. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 882.792905] env[62183]: DEBUG oslo_concurrency.lockutils [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] Acquiring lock "refresh_cache-71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.810299] env[62183]: DEBUG nova.compute.manager [req-da1b3568-6a03-4a15-b88f-0f41cfe74cee req-c276d0a5-9480-4256-bd93-7c068e274a55 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received event network-vif-plugged-e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.810717] env[62183]: DEBUG oslo_concurrency.lockutils [req-da1b3568-6a03-4a15-b88f-0f41cfe74cee req-c276d0a5-9480-4256-bd93-7c068e274a55 service nova] Acquiring lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.810844] env[62183]: DEBUG oslo_concurrency.lockutils [req-da1b3568-6a03-4a15-b88f-0f41cfe74cee req-c276d0a5-9480-4256-bd93-7c068e274a55 service nova] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.811034] env[62183]: DEBUG oslo_concurrency.lockutils [req-da1b3568-6a03-4a15-b88f-0f41cfe74cee req-c276d0a5-9480-4256-bd93-7c068e274a55 service nova] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.811224] env[62183]: DEBUG nova.compute.manager [req-da1b3568-6a03-4a15-b88f-0f41cfe74cee req-c276d0a5-9480-4256-bd93-7c068e274a55 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] No waiting events found dispatching network-vif-plugged-e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 882.811567] env[62183]: WARNING nova.compute.manager [req-da1b3568-6a03-4a15-b88f-0f41cfe74cee req-c276d0a5-9480-4256-bd93-7c068e274a55 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received unexpected event network-vif-plugged-e7836918-44a9-4db8-8daf-35c2032e5cdb for instance with vm_state active and task_state None. [ 882.944597] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 882.944791] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.341s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.993522] env[62183]: DEBUG nova.network.neutron [-] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.047179] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f84b48-c6bc-65a2-b5cc-d3fd96434dbd, 'name': SearchDatastore_Task, 'duration_secs': 0.080137} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.047517] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.047761] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.048018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.048178] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.048373] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.048775] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc5ad93d-0bfe-414d-934b-0f4d4f106b09 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.065879] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.066186] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.067023] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11f9eb7e-0dd1-4657-8432-f53b4da0dc48 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.075146] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 883.075146] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5269db5f-9e2f-69c3-daa8-e0ef1e512655" [ 883.075146] env[62183]: _type = "Task" [ 883.075146] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.084867] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5269db5f-9e2f-69c3-daa8-e0ef1e512655, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.122409] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387129, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.128611] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f620ac-76d4-423f-8595-7042be35371b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.150385] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd1f683-d1bf-4f43-b76e-b975be11e86d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.173289] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance '503786ca-dba4-43c1-9a25-9f1cbac9a6a4' progress to 67 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 883.196633] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.196848] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.196990] env[62183]: DEBUG nova.network.neutron [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.213906] env[62183]: DEBUG oslo_vmware.api [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387130, 'name': PowerOffVM_Task, 'duration_secs': 0.22302} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.214208] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 883.214405] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 883.214673] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-562d55d4-66cd-40d8-abc2-36e84c763ca5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.282292] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 883.282543] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 883.282726] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Deleting the datastore file [datastore1] 7465df88-c68c-49d3-9a91-6fff0d06957b {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.283016] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38cad4ad-c1ef-42b8-a363-3093ae05cbf7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.292789] env[62183]: DEBUG oslo_vmware.api [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for the task: (returnval){ [ 883.292789] env[62183]: value = "task-1387132" [ 883.292789] env[62183]: _type = "Task" [ 883.292789] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.303487] env[62183]: DEBUG oslo_vmware.api [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387132, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.304573] env[62183]: DEBUG nova.network.neutron [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 883.496827] env[62183]: INFO nova.compute.manager [-] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Took 1.21 seconds to deallocate network for instance. [ 883.506820] env[62183]: DEBUG nova.network.neutron [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Updating instance_info_cache with network_info: [{"id": "4e6d3d8d-6821-471a-bca5-29acccb74622", "address": "fa:16:3e:d2:4a:dd", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e6d3d8d-68", "ovs_interfaceid": "4e6d3d8d-6821-471a-bca5-29acccb74622", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.587723] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5269db5f-9e2f-69c3-daa8-e0ef1e512655, 'name': SearchDatastore_Task, 'duration_secs': 0.087713} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.588624] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e18f4e80-bbc1-422c-b812-822d73d53b45 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.596502] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 883.596502] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]526595a7-b150-ec3e-ae51-80ac90e1452e" [ 883.596502] env[62183]: _type = "Task" [ 883.596502] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.606769] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526595a7-b150-ec3e-ae51-80ac90e1452e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.622468] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387129, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.717956] env[62183]: DEBUG nova.network.neutron [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Port d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29 binding to destination host cpu-1 is already ACTIVE {{(pid=62183) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 883.734650] env[62183]: WARNING nova.network.neutron [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] ece15cf2-6921-4b76-9d7e-f313bcaa2f48 already exists in list: networks containing: ['ece15cf2-6921-4b76-9d7e-f313bcaa2f48']. ignoring it [ 883.808617] env[62183]: DEBUG oslo_vmware.api [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Task: {'id': task-1387132, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217779} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.809091] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.809091] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 883.809291] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.809474] env[62183]: INFO nova.compute.manager [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 883.809773] env[62183]: DEBUG oslo.service.loopingcall [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.809994] env[62183]: DEBUG nova.compute.manager [-] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.810107] env[62183]: DEBUG nova.network.neutron [-] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 884.003533] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.003827] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.004132] env[62183]: DEBUG nova.objects.instance [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lazy-loading 'resources' on Instance uuid 843d35ff-6871-4e2a-8e03-d7229a0d8246 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.009775] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "refresh_cache-71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.010151] env[62183]: DEBUG nova.compute.manager [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Instance network_info: |[{"id": "4e6d3d8d-6821-471a-bca5-29acccb74622", "address": "fa:16:3e:d2:4a:dd", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e6d3d8d-68", "ovs_interfaceid": "4e6d3d8d-6821-471a-bca5-29acccb74622", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 884.010458] env[62183]: DEBUG oslo_concurrency.lockutils [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] Acquired lock "refresh_cache-71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.010644] env[62183]: DEBUG nova.network.neutron [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Refreshing network info cache for port 4e6d3d8d-6821-471a-bca5-29acccb74622 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 884.011877] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:4a:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e6d3d8d-6821-471a-bca5-29acccb74622', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.020323] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Creating folder: Project (eda7e297c551407295b3c67fa0f1dbc0). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.021509] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09ec843c-808f-46cd-b6ef-b1fdd4fdbeca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.036388] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Created folder: Project (eda7e297c551407295b3c67fa0f1dbc0) in parent group-v294392. [ 884.036640] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Creating folder: Instances. Parent ref: group-v294467. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.036933] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a38d0a70-b340-42d3-8d3d-32d638a0786a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.051137] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Created folder: Instances in parent group-v294467. [ 884.051418] env[62183]: DEBUG oslo.service.loopingcall [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.051971] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.052236] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4adf08d3-2a3c-4eb2-b869-7515d9635fc5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.076533] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.076533] env[62183]: value = "task-1387135" [ 884.076533] env[62183]: _type = "Task" [ 884.076533] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.086427] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387135, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.090402] env[62183]: DEBUG nova.network.neutron [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e7836918-44a9-4db8-8daf-35c2032e5cdb", "address": "fa:16:3e:c8:56:ec", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7836918-44", "ovs_interfaceid": "e7836918-44a9-4db8-8daf-35c2032e5cdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.107691] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526595a7-b150-ec3e-ae51-80ac90e1452e, 'name': SearchDatastore_Task, 'duration_secs': 0.085825} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.107991] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.108281] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de/ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 884.108563] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8b4a521-5672-4921-83bd-9c25c76b2738 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.119688] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 884.119688] env[62183]: value = "task-1387136" [ 884.119688] env[62183]: _type = "Task" [ 884.119688] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.126687] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387129, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.133429] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387136, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.591626] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387135, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.593483] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.594180] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.594402] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.595240] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fde030-d309-4197-99cd-3dc97bb7d471 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.619469] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.619759] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.619923] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.620239] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.620427] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.620551] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.620760] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.620921] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.621105] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.621295] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.621472] env[62183]: DEBUG nova.virt.hardware [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.627951] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Reconfiguring VM to attach interface {{(pid=62183) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 884.636280] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-129b38e6-57a1-459f-a2f8-d1e1ee6a01fa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.651721] env[62183]: DEBUG nova.network.neutron [-] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.658918] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387136, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.663103] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387129, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.663428] env[62183]: DEBUG oslo_vmware.api [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 884.663428] env[62183]: value = "task-1387137" [ 884.663428] env[62183]: _type = "Task" [ 884.663428] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.677740] env[62183]: DEBUG oslo_vmware.api [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387137, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.747619] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.747871] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.748072] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.774204] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7d3258-c956-489f-a38d-fe586e19b9fe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.785312] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0034e0-6176-47be-a63e-d7d619d0e2d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.841916] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0958628e-ba6b-4a40-900e-7f8900cfc67f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.850216] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a189b92-fec7-4eaa-999e-d87e1988038b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.856893] env[62183]: DEBUG nova.compute.manager [req-ba2ea5dc-4f90-4d93-b2cf-54c501757cce req-96ae10bc-c65d-4299-9291-6b17a173ef5c service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received event network-changed-e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.857136] env[62183]: DEBUG nova.compute.manager [req-ba2ea5dc-4f90-4d93-b2cf-54c501757cce req-96ae10bc-c65d-4299-9291-6b17a173ef5c service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing instance network info cache due to event network-changed-e7836918-44a9-4db8-8daf-35c2032e5cdb. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 884.857386] env[62183]: DEBUG oslo_concurrency.lockutils [req-ba2ea5dc-4f90-4d93-b2cf-54c501757cce req-96ae10bc-c65d-4299-9291-6b17a173ef5c service nova] Acquiring lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.858114] env[62183]: DEBUG oslo_concurrency.lockutils [req-ba2ea5dc-4f90-4d93-b2cf-54c501757cce req-96ae10bc-c65d-4299-9291-6b17a173ef5c service nova] Acquired lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.858114] env[62183]: DEBUG nova.network.neutron [req-ba2ea5dc-4f90-4d93-b2cf-54c501757cce req-96ae10bc-c65d-4299-9291-6b17a173ef5c service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing network info cache for port e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 884.869357] env[62183]: DEBUG nova.compute.provider_tree [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.881588] env[62183]: DEBUG nova.network.neutron [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Updated VIF entry in instance network info cache for port 4e6d3d8d-6821-471a-bca5-29acccb74622. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 884.881963] env[62183]: DEBUG nova.network.neutron [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Updating instance_info_cache with network_info: [{"id": "4e6d3d8d-6821-471a-bca5-29acccb74622", "address": "fa:16:3e:d2:4a:dd", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e6d3d8d-68", "ovs_interfaceid": "4e6d3d8d-6821-471a-bca5-29acccb74622", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.088706] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387135, 'name': CreateVM_Task, 'duration_secs': 0.744677} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.088706] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.089431] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.089613] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.089999] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.090325] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3011507a-0560-4133-b7a5-89aefa728680 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.095647] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 885.095647] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52c4a744-ef26-1d97-8972-6e043abae11e" [ 885.095647] env[62183]: _type = "Task" [ 885.095647] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.104246] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c4a744-ef26-1d97-8972-6e043abae11e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.129452] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387129, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.367882} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.129769] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83/2c5496d6-0348-4ae3-9ab2-cefdc9e0ef83.vmdk to [datastore2] be021940-5fff-429a-9b93-4ad238cdff77/be021940-5fff-429a-9b93-4ad238cdff77.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 885.130814] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bcdad4-ada4-4942-8b3e-38072d581afe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.157035] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] be021940-5fff-429a-9b93-4ad238cdff77/be021940-5fff-429a-9b93-4ad238cdff77.vmdk or device None with type streamOptimized {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.161377] env[62183]: INFO nova.compute.manager [-] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Took 1.35 seconds to deallocate network for instance. [ 885.161704] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0e515de-e3a8-46ac-bdc8-5581d2c87188 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.189622] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387136, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.191718] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 885.191718] env[62183]: value = "task-1387138" [ 885.191718] env[62183]: _type = "Task" [ 885.191718] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.199079] env[62183]: DEBUG oslo_vmware.api [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387137, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.206074] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.373320] env[62183]: DEBUG nova.scheduler.client.report [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 885.384533] env[62183]: DEBUG oslo_concurrency.lockutils [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] Releasing lock "refresh_cache-71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.384918] env[62183]: DEBUG nova.compute.manager [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Received event network-vif-deleted-4266b7f9-08bc-469d-b4e8-a88b751a164e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.385164] env[62183]: INFO nova.compute.manager [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Neutron deleted interface 4266b7f9-08bc-469d-b4e8-a88b751a164e; detaching it from the instance and deleting it from the info cache [ 885.385394] env[62183]: DEBUG nova.network.neutron [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.579616] env[62183]: DEBUG nova.network.neutron [req-ba2ea5dc-4f90-4d93-b2cf-54c501757cce req-96ae10bc-c65d-4299-9291-6b17a173ef5c service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updated VIF entry in instance network info cache for port e7836918-44a9-4db8-8daf-35c2032e5cdb. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 885.580079] env[62183]: DEBUG nova.network.neutron [req-ba2ea5dc-4f90-4d93-b2cf-54c501757cce req-96ae10bc-c65d-4299-9291-6b17a173ef5c service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e7836918-44a9-4db8-8daf-35c2032e5cdb", "address": "fa:16:3e:c8:56:ec", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7836918-44", "ovs_interfaceid": "e7836918-44a9-4db8-8daf-35c2032e5cdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.607985] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c4a744-ef26-1d97-8972-6e043abae11e, 'name': SearchDatastore_Task, 'duration_secs': 0.054648} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.608299] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.608539] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.608775] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.608923] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.609120] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.609388] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-586a0ad3-fce4-4ece-86e2-0a7ae397b4ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.621236] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.621420] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.622131] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e21305f7-dc8d-4d74-986b-530b3921de46 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.627440] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 885.627440] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5254e67a-6afe-bac6-f03d-7024bb6d1ae7" [ 885.627440] env[62183]: _type = "Task" [ 885.627440] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.635069] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5254e67a-6afe-bac6-f03d-7024bb6d1ae7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.656142] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387136, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.278778} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.656416] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de/ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 885.656632] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.657409] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84e424e8-a593-48a6-9787-0c631082ca36 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.663048] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 885.663048] env[62183]: value = "task-1387139" [ 885.663048] env[62183]: _type = "Task" [ 885.663048] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.670299] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387139, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.687139] env[62183]: DEBUG oslo_vmware.api [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387137, 'name': ReconfigVM_Task, 'duration_secs': 0.900136} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.687615] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.687840] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Reconfigured VM to attach interface {{(pid=62183) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 885.692338] env[62183]: DEBUG oslo_concurrency.lockutils [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.700058] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387138, 'name': ReconfigVM_Task, 'duration_secs': 0.370678} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.700317] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Reconfigured VM instance instance-0000004e to attach disk [datastore2] be021940-5fff-429a-9b93-4ad238cdff77/be021940-5fff-429a-9b93-4ad238cdff77.vmdk or device None with type streamOptimized {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.700928] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93371dca-eaae-46c4-8b67-9ab5742c64e7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.708745] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 885.708745] env[62183]: value = "task-1387140" [ 885.708745] env[62183]: _type = "Task" [ 885.708745] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.720842] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387140, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.799851] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.800067] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.800260] env[62183]: DEBUG nova.network.neutron [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 885.878455] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.874s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.880903] env[62183]: DEBUG oslo_concurrency.lockutils [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.189s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.881215] env[62183]: DEBUG nova.objects.instance [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lazy-loading 'resources' on Instance uuid 7465df88-c68c-49d3-9a91-6fff0d06957b {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.888460] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c756342-8aec-460a-971f-0015e9b9b414 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.897272] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abddd371-0686-4478-b5df-f32e4d2e5e4b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.908449] env[62183]: INFO nova.scheduler.client.report [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted allocations for instance 843d35ff-6871-4e2a-8e03-d7229a0d8246 [ 885.930239] env[62183]: DEBUG nova.compute.manager [req-7b68042b-dbc8-4bb3-8387-74e29f17f9cc req-edfdc9c2-2db5-4951-afa0-128e1f9c3972 service nova] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Detach interface failed, port_id=4266b7f9-08bc-469d-b4e8-a88b751a164e, reason: Instance 843d35ff-6871-4e2a-8e03-d7229a0d8246 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 886.082656] env[62183]: DEBUG oslo_concurrency.lockutils [req-ba2ea5dc-4f90-4d93-b2cf-54c501757cce req-96ae10bc-c65d-4299-9291-6b17a173ef5c service nova] Releasing lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.082897] env[62183]: DEBUG nova.compute.manager [req-ba2ea5dc-4f90-4d93-b2cf-54c501757cce req-96ae10bc-c65d-4299-9291-6b17a173ef5c service nova] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Received event network-vif-deleted-dccdd9b2-d8fa-435b-86d8-ec505124174c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.138023] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5254e67a-6afe-bac6-f03d-7024bb6d1ae7, 'name': SearchDatastore_Task, 'duration_secs': 0.028749} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.138803] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f45c5877-9894-4d31-8bc6-9def6085c2a7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.143675] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 886.143675] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]524ad7de-1330-89f7-1690-dacadd4dc24f" [ 886.143675] env[62183]: _type = "Task" [ 886.143675] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.151050] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]524ad7de-1330-89f7-1690-dacadd4dc24f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.170186] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387139, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079684} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.170503] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.171214] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c4bf3c-176f-4991-ad04-0c12b2ae771c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.192484] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de/ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.194216] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5395b5c1-8a2f-4a17-9129-4dfea8f39942 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "interface-ac555ffc-ce4e-4650-97fd-c26a3246fe4b-e7836918-44a9-4db8-8daf-35c2032e5cdb" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.643s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.194575] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fec11370-e807-4d11-9daa-03a51c1186d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.214980] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 886.214980] env[62183]: value = "task-1387141" [ 886.214980] env[62183]: _type = "Task" [ 886.214980] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.217833] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387140, 'name': Rename_Task, 'duration_secs': 0.131495} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.220578] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.220818] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6d48c23-6f63-4e00-933f-c4b560a564c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.227244] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387141, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.228359] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 886.228359] env[62183]: value = "task-1387142" [ 886.228359] env[62183]: _type = "Task" [ 886.228359] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.235383] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387142, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.415590] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5c0bd094-ce49-4ce4-abd9-471db6c9b83e tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "843d35ff-6871-4e2a-8e03-d7229a0d8246" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.252s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.558535] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f31138-5b35-408b-a1ea-4cbd4f801007 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.566336] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a59f7a2-c9fd-45dc-b0b5-798aef024849 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.599850] env[62183]: DEBUG nova.network.neutron [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance_info_cache with network_info: [{"id": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "address": "fa:16:3e:2f:6f:39", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2f9a654-1f", "ovs_interfaceid": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.601500] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9c7ce7-e994-4c89-bb8d-7d7b4adcff6b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.609341] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1baa30db-f206-4757-9161-c15960ffc553 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.623539] env[62183]: DEBUG nova.compute.provider_tree [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.654153] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]524ad7de-1330-89f7-1690-dacadd4dc24f, 'name': SearchDatastore_Task, 'duration_secs': 0.017075} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.654455] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.654749] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e/71fb3a2c-efd3-4f72-9997-136ee2ee4f6e.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.655039] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-257d03a6-f6f3-4300-807a-a9d6a735b2ab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.661393] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 886.661393] env[62183]: value = "task-1387143" [ 886.661393] env[62183]: _type = "Task" [ 886.661393] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.670299] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387143, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.727594] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387141, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.737170] env[62183]: DEBUG oslo_vmware.api [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387142, 'name': PowerOnVM_Task, 'duration_secs': 0.458115} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.737457] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.737660] env[62183]: INFO nova.compute.manager [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Took 13.14 seconds to spawn the instance on the hypervisor. [ 886.737852] env[62183]: DEBUG nova.compute.manager [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.738633] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ce8e66-ec16-4937-9272-bf6ba6f37ec1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.106042] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.127294] env[62183]: DEBUG nova.scheduler.client.report [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 887.175978] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387143, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.228832] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387141, 'name': ReconfigVM_Task, 'duration_secs': 0.585336} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.228832] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Reconfigured VM instance instance-0000004f to attach disk [datastore2] ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de/ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.229282] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ac78df4-12c4-45c2-a270-a3a195380045 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.237018] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 887.237018] env[62183]: value = "task-1387144" [ 887.237018] env[62183]: _type = "Task" [ 887.237018] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.261290] env[62183]: INFO nova.compute.manager [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Took 24.27 seconds to build instance. [ 887.263180] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387144, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.377994] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.378261] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.629369] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea74b8d5-3308-49c6-9e14-136385ef0794 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.632618] env[62183]: DEBUG oslo_concurrency.lockutils [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.752s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.652066] env[62183]: INFO nova.scheduler.client.report [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Deleted allocations for instance 7465df88-c68c-49d3-9a91-6fff0d06957b [ 887.653556] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64c6b96-baaf-4d86-9218-9d4426467b7a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.664986] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance '503786ca-dba4-43c1-9a25-9f1cbac9a6a4' progress to 83 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 887.677022] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387143, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585449} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.677022] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e/71fb3a2c-efd3-4f72-9997-136ee2ee4f6e.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.677022] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.677267] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51412c24-e138-4144-8572-d7fd574738cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.683821] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 887.683821] env[62183]: value = "task-1387145" [ 887.683821] env[62183]: _type = "Task" [ 887.683821] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.691821] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387145, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.747422] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387144, 'name': Rename_Task, 'duration_secs': 0.142437} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.747759] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 887.748063] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e188b4b7-0387-43ea-9814-b88dfb453fdb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.754412] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "interface-ac555ffc-ce4e-4650-97fd-c26a3246fe4b-e7836918-44a9-4db8-8daf-35c2032e5cdb" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.754735] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "interface-ac555ffc-ce4e-4650-97fd-c26a3246fe4b-e7836918-44a9-4db8-8daf-35c2032e5cdb" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.759050] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 887.759050] env[62183]: value = "task-1387146" [ 887.759050] env[62183]: _type = "Task" [ 887.759050] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.763216] env[62183]: DEBUG oslo_concurrency.lockutils [None req-458df5dd-cfd6-462b-ae53-9d8fd93094cf tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "be021940-5fff-429a-9b93-4ad238cdff77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.777s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.766921] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387146, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.881169] env[62183]: DEBUG nova.compute.manager [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.019499] env[62183]: DEBUG oslo_concurrency.lockutils [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "be021940-5fff-429a-9b93-4ad238cdff77" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.019787] env[62183]: DEBUG oslo_concurrency.lockutils [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "be021940-5fff-429a-9b93-4ad238cdff77" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.019997] env[62183]: DEBUG oslo_concurrency.lockutils [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "be021940-5fff-429a-9b93-4ad238cdff77-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.020199] env[62183]: DEBUG oslo_concurrency.lockutils [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "be021940-5fff-429a-9b93-4ad238cdff77-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.020420] env[62183]: DEBUG oslo_concurrency.lockutils [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "be021940-5fff-429a-9b93-4ad238cdff77-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.022614] env[62183]: INFO nova.compute.manager [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Terminating instance [ 888.024470] env[62183]: DEBUG nova.compute.manager [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 888.025175] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 888.025459] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-098200f7-114a-4956-b513-1bbed061888b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.034408] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 888.034699] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-625deda3-7169-4017-9606-d4a29bda586c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.042314] env[62183]: DEBUG oslo_vmware.api [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 888.042314] env[62183]: value = "task-1387147" [ 888.042314] env[62183]: _type = "Task" [ 888.042314] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.052091] env[62183]: DEBUG oslo_vmware.api [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387147, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.163714] env[62183]: DEBUG oslo_concurrency.lockutils [None req-550253a6-53f9-4042-94b4-d6516fd01b2e tempest-FloatingIPsAssociationTestJSON-2079894705 tempest-FloatingIPsAssociationTestJSON-2079894705-project-member] Lock "7465df88-c68c-49d3-9a91-6fff0d06957b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.486s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.173466] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.173773] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37617118-d733-451c-b213-3988704ceba8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.187039] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 888.187039] env[62183]: value = "task-1387148" [ 888.187039] env[62183]: _type = "Task" [ 888.187039] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.198052] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387145, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07221} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.201561] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.201980] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387148, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.202802] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1e9f11-cc2d-4ecd-9a7c-1fca3cea6907 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.234012] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e/71fb3a2c-efd3-4f72-9997-136ee2ee4f6e.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.234297] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c105a32b-0309-4cb3-b3a3-e4e293a00911 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.259260] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.259629] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.263664] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37bf41b-af25-4e34-b040-3e62ab20526e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.268979] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 888.268979] env[62183]: value = "task-1387149" [ 888.268979] env[62183]: _type = "Task" [ 888.268979] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.290909] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387146, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.292347] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5319b54b-ee90-4d0c-b222-772c4ff1719c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.298437] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.322592] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Reconfiguring VM to detach interface {{(pid=62183) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 888.323279] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af9a1398-b543-49ea-89d5-559c59166bfd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.344509] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 888.344509] env[62183]: value = "task-1387150" [ 888.344509] env[62183]: _type = "Task" [ 888.344509] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.354821] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.406427] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.407275] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.408491] env[62183]: INFO nova.compute.claims [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.553782] env[62183]: DEBUG oslo_vmware.api [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387147, 'name': PowerOffVM_Task, 'duration_secs': 0.216327} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.554226] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.554338] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 888.554708] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97b864aa-3d37-43cb-89d5-f00102bffaaa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.689127] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 888.689127] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 888.689127] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleting the datastore file [datastore2] be021940-5fff-429a-9b93-4ad238cdff77 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.689127] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-464d163a-4a61-4f8f-9020-00413115bd4b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.700550] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387148, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.702529] env[62183]: DEBUG oslo_vmware.api [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 888.702529] env[62183]: value = "task-1387152" [ 888.702529] env[62183]: _type = "Task" [ 888.702529] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.712326] env[62183]: DEBUG oslo_vmware.api [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387152, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.774165] env[62183]: DEBUG oslo_vmware.api [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387146, 'name': PowerOnVM_Task, 'duration_secs': 0.526279} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.775052] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 888.775606] env[62183]: INFO nova.compute.manager [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Took 10.40 seconds to spawn the instance on the hypervisor. [ 888.775993] env[62183]: DEBUG nova.compute.manager [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.777118] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe79708-db26-46cc-9a08-ce7861d57fdc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.788717] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387149, 'name': ReconfigVM_Task, 'duration_secs': 0.290361} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.790724] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e/71fb3a2c-efd3-4f72-9997-136ee2ee4f6e.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.800350] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65244b5c-425e-4a1a-be09-85d9d6c934da {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.808852] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 888.808852] env[62183]: value = "task-1387153" [ 888.808852] env[62183]: _type = "Task" [ 888.808852] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.818594] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387153, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.855601] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.089463] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.089764] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.089956] env[62183]: DEBUG nova.compute.manager [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.090913] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d0ac61-b504-4b04-acc5-1dcd864a994a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.099408] env[62183]: DEBUG nova.compute.manager [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62183) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 889.099883] env[62183]: DEBUG nova.objects.instance [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'flavor' on Instance uuid a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.199048] env[62183]: DEBUG oslo_vmware.api [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387148, 'name': PowerOnVM_Task, 'duration_secs': 0.60771} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.199427] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.199518] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4ab79f-e10a-406f-bf16-6d2dd8605452 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance '503786ca-dba4-43c1-9a25-9f1cbac9a6a4' progress to 100 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 889.212989] env[62183]: DEBUG oslo_vmware.api [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387152, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152649} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.213276] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.213462] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 889.213648] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 889.213864] env[62183]: INFO nova.compute.manager [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Took 1.19 seconds to destroy the instance on the hypervisor. [ 889.214102] env[62183]: DEBUG oslo.service.loopingcall [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.214299] env[62183]: DEBUG nova.compute.manager [-] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.214395] env[62183]: DEBUG nova.network.neutron [-] [instance: be021940-5fff-429a-9b93-4ad238cdff77] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 889.314581] env[62183]: INFO nova.compute.manager [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Took 20.48 seconds to build instance. [ 889.321892] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387153, 'name': Rename_Task, 'duration_secs': 0.242085} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.322443] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 889.323083] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55bc93b5-2518-4e64-966e-7189b39fcc2e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.335024] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 889.335024] env[62183]: value = "task-1387154" [ 889.335024] env[62183]: _type = "Task" [ 889.335024] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.343114] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387154, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.354537] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.557008] env[62183]: DEBUG nova.compute.manager [req-26a0ceca-f95b-4c01-a282-ab0385473209 req-5f1d0a2d-b077-4cf5-8dad-b530d252a6cd service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Received event network-vif-deleted-1d3eaacd-3e0b-4985-baa2-11f90dba863c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.557253] env[62183]: INFO nova.compute.manager [req-26a0ceca-f95b-4c01-a282-ab0385473209 req-5f1d0a2d-b077-4cf5-8dad-b530d252a6cd service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Neutron deleted interface 1d3eaacd-3e0b-4985-baa2-11f90dba863c; detaching it from the instance and deleting it from the info cache [ 889.557432] env[62183]: DEBUG nova.network.neutron [req-26a0ceca-f95b-4c01-a282-ab0385473209 req-5f1d0a2d-b077-4cf5-8dad-b530d252a6cd service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.604674] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.604941] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9c8f541-080f-4d8a-8e00-45a2f8a05abb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.613251] env[62183]: DEBUG oslo_vmware.api [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 889.613251] env[62183]: value = "task-1387155" [ 889.613251] env[62183]: _type = "Task" [ 889.613251] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.624360] env[62183]: DEBUG oslo_vmware.api [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387155, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.650716] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef080d3-fcdc-4a23-adfb-3c9a373f3c2f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.659552] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72f7d01-bf7a-4326-952b-a0c9e4fb0534 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.693068] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e540e8-053f-4df2-98e1-97bc595edfd4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.702065] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8349b41f-d8a1-46c0-9a15-4de6ed2dd6d8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.719049] env[62183]: DEBUG nova.compute.provider_tree [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.817354] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92c10457-2152-458a-ae8b-2f3ee6ddbe85 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.988s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.846962] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387154, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.856816] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.021735] env[62183]: DEBUG nova.network.neutron [-] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.062742] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-516e91b5-d06f-4b91-b14c-619a1d3cf00d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.074133] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20235859-bfb6-4a2c-a2b0-67771cbf2f5e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.112735] env[62183]: DEBUG nova.compute.manager [req-26a0ceca-f95b-4c01-a282-ab0385473209 req-5f1d0a2d-b077-4cf5-8dad-b530d252a6cd service nova] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Detach interface failed, port_id=1d3eaacd-3e0b-4985-baa2-11f90dba863c, reason: Instance be021940-5fff-429a-9b93-4ad238cdff77 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 890.125065] env[62183]: DEBUG oslo_vmware.api [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387155, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.225795] env[62183]: DEBUG nova.scheduler.client.report [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 890.313377] env[62183]: DEBUG nova.compute.manager [req-9386ee30-5f42-427d-bff7-443c6ef9f7cb req-8ac3f9a5-cc0b-45bb-82be-d714972de349 service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Received event network-changed-8d0bfbea-f6ad-41ee-a0f1-4ccad145901f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.313453] env[62183]: DEBUG nova.compute.manager [req-9386ee30-5f42-427d-bff7-443c6ef9f7cb req-8ac3f9a5-cc0b-45bb-82be-d714972de349 service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Refreshing instance network info cache due to event network-changed-8d0bfbea-f6ad-41ee-a0f1-4ccad145901f. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 890.313697] env[62183]: DEBUG oslo_concurrency.lockutils [req-9386ee30-5f42-427d-bff7-443c6ef9f7cb req-8ac3f9a5-cc0b-45bb-82be-d714972de349 service nova] Acquiring lock "refresh_cache-ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.313819] env[62183]: DEBUG oslo_concurrency.lockutils [req-9386ee30-5f42-427d-bff7-443c6ef9f7cb req-8ac3f9a5-cc0b-45bb-82be-d714972de349 service nova] Acquired lock "refresh_cache-ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.314024] env[62183]: DEBUG nova.network.neutron [req-9386ee30-5f42-427d-bff7-443c6ef9f7cb req-8ac3f9a5-cc0b-45bb-82be-d714972de349 service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Refreshing network info cache for port 8d0bfbea-f6ad-41ee-a0f1-4ccad145901f {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.347324] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387154, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.356714] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.525939] env[62183]: INFO nova.compute.manager [-] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Took 1.31 seconds to deallocate network for instance. [ 890.625777] env[62183]: DEBUG oslo_vmware.api [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387155, 'name': PowerOffVM_Task, 'duration_secs': 0.620105} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.626063] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 890.626240] env[62183]: DEBUG nova.compute.manager [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.627069] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8fbbd2-6ef5-4bf3-817a-57a546572312 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.735944] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.736395] env[62183]: DEBUG nova.compute.manager [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 890.847253] env[62183]: DEBUG oslo_vmware.api [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387154, 'name': PowerOnVM_Task, 'duration_secs': 1.173748} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.849685] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 890.849931] env[62183]: INFO nova.compute.manager [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Took 10.11 seconds to spawn the instance on the hypervisor. [ 890.850135] env[62183]: DEBUG nova.compute.manager [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.853529] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee5210f0-04e6-4ce0-a618-e4d8dfa9639e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.864164] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.032286] env[62183]: DEBUG oslo_concurrency.lockutils [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.032674] env[62183]: DEBUG oslo_concurrency.lockutils [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.032910] env[62183]: DEBUG nova.objects.instance [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lazy-loading 'resources' on Instance uuid be021940-5fff-429a-9b93-4ad238cdff77 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.800704] env[62183]: DEBUG oslo_concurrency.lockutils [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.801039] env[62183]: DEBUG oslo_concurrency.lockutils [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.801141] env[62183]: DEBUG nova.compute.manager [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Going to confirm migration 1 {{(pid=62183) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 891.806856] env[62183]: DEBUG nova.compute.utils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.812383] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a8fcfdd5-340b-45ac-b0a9-d71bf3d5a660 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.723s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.815591] env[62183]: DEBUG nova.compute.manager [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.815758] env[62183]: DEBUG nova.network.neutron [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 891.829164] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.834875] env[62183]: INFO nova.compute.manager [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Took 17.52 seconds to build instance. [ 891.871682] env[62183]: DEBUG nova.policy [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59edae5aed2e4ccbb980bc0973890baf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '594f9c41496146c68921dad4c8d99ec4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 891.971849] env[62183]: DEBUG nova.network.neutron [req-9386ee30-5f42-427d-bff7-443c6ef9f7cb req-8ac3f9a5-cc0b-45bb-82be-d714972de349 service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Updated VIF entry in instance network info cache for port 8d0bfbea-f6ad-41ee-a0f1-4ccad145901f. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 891.971849] env[62183]: DEBUG nova.network.neutron [req-9386ee30-5f42-427d-bff7-443c6ef9f7cb req-8ac3f9a5-cc0b-45bb-82be-d714972de349 service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Updating instance_info_cache with network_info: [{"id": "8d0bfbea-f6ad-41ee-a0f1-4ccad145901f", "address": "fa:16:3e:07:ab:a2", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8d0bfbea-f6", "ovs_interfaceid": "8d0bfbea-f6ad-41ee-a0f1-4ccad145901f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.072092] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5932cde-d3ba-44fd-9253-0e2b9a424e94 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.080806] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f4b06a-0d3a-46ab-8c31-7aabbe51ad76 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.113929] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ccc1c1f-4947-4d61-b025-cb31e2a31040 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.123194] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a4fb20-b920-4b4c-84e5-e16a34b30ca0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.140254] env[62183]: DEBUG nova.compute.provider_tree [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.146352] env[62183]: DEBUG nova.network.neutron [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Successfully created port: 9c452a0f-d3ff-4e17-92b6-a02b07bc1de2 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.311968] env[62183]: DEBUG nova.compute.manager [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.320904] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.335939] env[62183]: DEBUG nova.objects.instance [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'flavor' on Instance uuid a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.340503] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a69c4416-1408-482c-a412-b98ca3745f74 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.037s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.453412] env[62183]: DEBUG oslo_concurrency.lockutils [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.453585] env[62183]: DEBUG oslo_concurrency.lockutils [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.453763] env[62183]: DEBUG nova.network.neutron [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.454020] env[62183]: DEBUG nova.objects.instance [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lazy-loading 'info_cache' on Instance uuid 503786ca-dba4-43c1-9a25-9f1cbac9a6a4 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.474801] env[62183]: DEBUG oslo_concurrency.lockutils [req-9386ee30-5f42-427d-bff7-443c6ef9f7cb req-8ac3f9a5-cc0b-45bb-82be-d714972de349 service nova] Releasing lock "refresh_cache-ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.643636] env[62183]: DEBUG nova.scheduler.client.report [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 892.817407] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.841888] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.842083] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.842265] env[62183]: DEBUG nova.network.neutron [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.842518] env[62183]: DEBUG nova.objects.instance [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'info_cache' on Instance uuid a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.002606] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.003084] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.148041] env[62183]: DEBUG oslo_concurrency.lockutils [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.115s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.166047] env[62183]: INFO nova.scheduler.client.report [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleted allocations for instance be021940-5fff-429a-9b93-4ad238cdff77 [ 893.317920] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.327331] env[62183]: DEBUG nova.compute.manager [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.345645] env[62183]: DEBUG nova.objects.base [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 893.356041] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.356041] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.356277] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.356417] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.356585] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.356779] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.357047] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.357233] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.357411] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.357594] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.357777] env[62183]: DEBUG nova.virt.hardware [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.358715] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db177330-447e-4a8f-a83c-65bb6b1c7fca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.369884] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269354f4-6d4f-461f-99bb-1ef496114b4c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.506322] env[62183]: DEBUG nova.compute.manager [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 893.679112] env[62183]: DEBUG oslo_concurrency.lockutils [None req-03965b00-fe93-4800-9145-2aa7c61a04d3 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "be021940-5fff-429a-9b93-4ad238cdff77" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.659s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.758976] env[62183]: DEBUG nova.compute.manager [req-842c53da-975d-4b75-a1f9-2f921c26611c req-a4aa5cc1-1711-44c4-bbe4-70f19663bfd4 service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Received event network-vif-plugged-9c452a0f-d3ff-4e17-92b6-a02b07bc1de2 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.759338] env[62183]: DEBUG oslo_concurrency.lockutils [req-842c53da-975d-4b75-a1f9-2f921c26611c req-a4aa5cc1-1711-44c4-bbe4-70f19663bfd4 service nova] Acquiring lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.759636] env[62183]: DEBUG oslo_concurrency.lockutils [req-842c53da-975d-4b75-a1f9-2f921c26611c req-a4aa5cc1-1711-44c4-bbe4-70f19663bfd4 service nova] Lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.759879] env[62183]: DEBUG oslo_concurrency.lockutils [req-842c53da-975d-4b75-a1f9-2f921c26611c req-a4aa5cc1-1711-44c4-bbe4-70f19663bfd4 service nova] Lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.760148] env[62183]: DEBUG nova.compute.manager [req-842c53da-975d-4b75-a1f9-2f921c26611c req-a4aa5cc1-1711-44c4-bbe4-70f19663bfd4 service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] No waiting events found dispatching network-vif-plugged-9c452a0f-d3ff-4e17-92b6-a02b07bc1de2 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 893.760403] env[62183]: WARNING nova.compute.manager [req-842c53da-975d-4b75-a1f9-2f921c26611c req-a4aa5cc1-1711-44c4-bbe4-70f19663bfd4 service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Received unexpected event network-vif-plugged-9c452a0f-d3ff-4e17-92b6-a02b07bc1de2 for instance with vm_state building and task_state spawning. [ 893.762719] env[62183]: DEBUG nova.network.neutron [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance_info_cache with network_info: [{"id": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "address": "fa:16:3e:2f:6f:39", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2f9a654-1f", "ovs_interfaceid": "d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.819491] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.850181] env[62183]: DEBUG nova.network.neutron [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Successfully updated port: 9c452a0f-d3ff-4e17-92b6-a02b07bc1de2 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.030958] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.031243] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.032762] env[62183]: INFO nova.compute.claims [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.169294] env[62183]: DEBUG nova.network.neutron [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.265396] env[62183]: DEBUG oslo_concurrency.lockutils [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-503786ca-dba4-43c1-9a25-9f1cbac9a6a4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.265759] env[62183]: DEBUG nova.objects.instance [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lazy-loading 'migration_context' on Instance uuid 503786ca-dba4-43c1-9a25-9f1cbac9a6a4 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.319510] env[62183]: DEBUG oslo_vmware.api [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387150, 'name': ReconfigVM_Task, 'duration_secs': 5.839552} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.319510] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.319685] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Reconfigured VM to detach interface {{(pid=62183) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 894.354220] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-14ea1b33-ffbd-4c48-84ea-443c10d8317d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.354341] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-14ea1b33-ffbd-4c48-84ea-443c10d8317d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.354473] env[62183]: DEBUG nova.network.neutron [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.673796] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.768076] env[62183]: DEBUG nova.objects.base [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Object Instance<503786ca-dba4-43c1-9a25-9f1cbac9a6a4> lazy-loaded attributes: info_cache,migration_context {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 894.769171] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca5bad5-8e7a-4624-acaf-da4c741a9848 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.790338] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcf3a2a5-e0b0-4222-9e60-a86d57db234b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.796298] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.796533] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.796756] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.796944] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.797126] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.798981] env[62183]: DEBUG oslo_vmware.api [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 894.798981] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]526e64db-4cf6-a75d-9579-abeee3c51f4c" [ 894.798981] env[62183]: _type = "Task" [ 894.798981] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.799490] env[62183]: INFO nova.compute.manager [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Terminating instance [ 894.804218] env[62183]: DEBUG nova.compute.manager [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 894.804418] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.805423] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac11616-31fe-4df1-9f40-96a77f74e56e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.815242] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.815745] env[62183]: DEBUG oslo_vmware.api [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526e64db-4cf6-a75d-9579-abeee3c51f4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.815961] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89886506-865f-415c-af2e-16bd2940f8de {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.821928] env[62183]: DEBUG oslo_vmware.api [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 894.821928] env[62183]: value = "task-1387156" [ 894.821928] env[62183]: _type = "Task" [ 894.821928] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.829854] env[62183]: DEBUG oslo_vmware.api [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387156, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.018884] env[62183]: DEBUG nova.network.neutron [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.167730] env[62183]: DEBUG nova.network.neutron [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Updating instance_info_cache with network_info: [{"id": "9c452a0f-d3ff-4e17-92b6-a02b07bc1de2", "address": "fa:16:3e:ff:11:3f", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c452a0f-d3", "ovs_interfaceid": "9c452a0f-d3ff-4e17-92b6-a02b07bc1de2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.176983] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.177295] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-733e748f-fffc-436b-8652-354b385616df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.186600] env[62183]: DEBUG oslo_vmware.api [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 895.186600] env[62183]: value = "task-1387157" [ 895.186600] env[62183]: _type = "Task" [ 895.186600] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.194310] env[62183]: DEBUG oslo_vmware.api [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387157, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.226157] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681895e9-bdae-4a6d-8e52-43913d677c91 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.233368] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2582da9e-f0b6-4b59-9cfa-8ddf67656513 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.266323] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536d901b-da6b-4989-b472-60c5e721c0cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.275196] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a6e029-ad8e-4a60-ad30-7c02bf3eed59 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.289954] env[62183]: DEBUG nova.compute.provider_tree [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.308928] env[62183]: DEBUG oslo_vmware.api [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526e64db-4cf6-a75d-9579-abeee3c51f4c, 'name': SearchDatastore_Task, 'duration_secs': 0.211698} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.309243] env[62183]: DEBUG oslo_concurrency.lockutils [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.331987] env[62183]: DEBUG oslo_vmware.api [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387156, 'name': PowerOffVM_Task, 'duration_secs': 0.31902} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.332288] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.332460] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.332712] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82e3ce38-7f54-4175-a1c5-4e9c61134721 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.362223] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "46aa1062-3fcd-4581-8647-32e1372e9491" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.362540] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "46aa1062-3fcd-4581-8647-32e1372e9491" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.393052] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.393309] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.393485] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleting the datastore file [datastore1] 850d40d8-565a-49a2-a27f-3de2a8dc7e30 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.393755] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c1763ba-43d9-45b6-8aca-52ede233498e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.400134] env[62183]: DEBUG oslo_vmware.api [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 895.400134] env[62183]: value = "task-1387159" [ 895.400134] env[62183]: _type = "Task" [ 895.400134] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.408230] env[62183]: DEBUG oslo_vmware.api [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.670679] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-14ea1b33-ffbd-4c48-84ea-443c10d8317d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.670989] env[62183]: DEBUG nova.compute.manager [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Instance network_info: |[{"id": "9c452a0f-d3ff-4e17-92b6-a02b07bc1de2", "address": "fa:16:3e:ff:11:3f", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c452a0f-d3", "ovs_interfaceid": "9c452a0f-d3ff-4e17-92b6-a02b07bc1de2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.671460] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:11:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '496ac502-bfc4-4324-8332-cac473eb7cc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c452a0f-d3ff-4e17-92b6-a02b07bc1de2', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.678945] env[62183]: DEBUG oslo.service.loopingcall [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.679183] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 895.679408] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37606343-eae3-418b-a41c-958c192567f1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.701967] env[62183]: DEBUG oslo_vmware.api [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387157, 'name': PowerOnVM_Task, 'duration_secs': 0.432836} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.703535] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.703535] env[62183]: DEBUG nova.compute.manager [None req-c407aab9-2740-40fe-89e8-e34890d8677f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.703732] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.703732] env[62183]: value = "task-1387160" [ 895.703732] env[62183]: _type = "Task" [ 895.703732] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.704363] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3d6feb-748e-4536-b2e6-8c31f6d91cfb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.718344] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387160, 'name': CreateVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.771729] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.771951] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.772167] env[62183]: DEBUG nova.network.neutron [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 895.792546] env[62183]: DEBUG nova.compute.manager [req-9252b6fd-2845-4817-947b-0f7a5759ab52 req-ac7220f2-b310-49b1-a6f8-6c2350ec7c9c service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Received event network-changed-9c452a0f-d3ff-4e17-92b6-a02b07bc1de2 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.792760] env[62183]: DEBUG nova.compute.manager [req-9252b6fd-2845-4817-947b-0f7a5759ab52 req-ac7220f2-b310-49b1-a6f8-6c2350ec7c9c service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Refreshing instance network info cache due to event network-changed-9c452a0f-d3ff-4e17-92b6-a02b07bc1de2. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 895.792974] env[62183]: DEBUG oslo_concurrency.lockutils [req-9252b6fd-2845-4817-947b-0f7a5759ab52 req-ac7220f2-b310-49b1-a6f8-6c2350ec7c9c service nova] Acquiring lock "refresh_cache-14ea1b33-ffbd-4c48-84ea-443c10d8317d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.793140] env[62183]: DEBUG oslo_concurrency.lockutils [req-9252b6fd-2845-4817-947b-0f7a5759ab52 req-ac7220f2-b310-49b1-a6f8-6c2350ec7c9c service nova] Acquired lock "refresh_cache-14ea1b33-ffbd-4c48-84ea-443c10d8317d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.793302] env[62183]: DEBUG nova.network.neutron [req-9252b6fd-2845-4817-947b-0f7a5759ab52 req-ac7220f2-b310-49b1-a6f8-6c2350ec7c9c service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Refreshing network info cache for port 9c452a0f-d3ff-4e17-92b6-a02b07bc1de2 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 895.795299] env[62183]: DEBUG nova.scheduler.client.report [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 895.868269] env[62183]: DEBUG nova.compute.manager [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 895.909762] env[62183]: DEBUG oslo_vmware.api [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387159, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133795} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.911456] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.911456] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.911456] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.911663] env[62183]: INFO nova.compute.manager [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Took 1.11 seconds to destroy the instance on the hypervisor. [ 895.911820] env[62183]: DEBUG oslo.service.loopingcall [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.912316] env[62183]: DEBUG nova.compute.manager [-] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.912431] env[62183]: DEBUG nova.network.neutron [-] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 896.218398] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387160, 'name': CreateVM_Task, 'duration_secs': 0.363013} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.218563] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 896.219317] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.219491] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.219827] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.222132] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06bb758f-3782-4ddb-87ab-936cfc5da61e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.227958] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 896.227958] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5200c01c-a228-b093-37ab-a486b1cce79c" [ 896.227958] env[62183]: _type = "Task" [ 896.227958] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.238299] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5200c01c-a228-b093-37ab-a486b1cce79c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.303024] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.303024] env[62183]: DEBUG nova.compute.manager [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.305413] env[62183]: DEBUG oslo_concurrency.lockutils [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.996s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.397730] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.608184] env[62183]: INFO nova.network.neutron [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Port e7836918-44a9-4db8-8daf-35c2032e5cdb from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 896.608511] env[62183]: DEBUG nova.network.neutron [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.710622] env[62183]: DEBUG nova.network.neutron [req-9252b6fd-2845-4817-947b-0f7a5759ab52 req-ac7220f2-b310-49b1-a6f8-6c2350ec7c9c service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Updated VIF entry in instance network info cache for port 9c452a0f-d3ff-4e17-92b6-a02b07bc1de2. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 896.711176] env[62183]: DEBUG nova.network.neutron [req-9252b6fd-2845-4817-947b-0f7a5759ab52 req-ac7220f2-b310-49b1-a6f8-6c2350ec7c9c service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Updating instance_info_cache with network_info: [{"id": "9c452a0f-d3ff-4e17-92b6-a02b07bc1de2", "address": "fa:16:3e:ff:11:3f", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c452a0f-d3", "ovs_interfaceid": "9c452a0f-d3ff-4e17-92b6-a02b07bc1de2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.714190] env[62183]: DEBUG nova.network.neutron [-] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.738636] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5200c01c-a228-b093-37ab-a486b1cce79c, 'name': SearchDatastore_Task, 'duration_secs': 0.012715} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.738987] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.739239] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.739466] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.739615] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.739796] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.740065] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0661eefd-50ee-4c7e-bb67-d61380544b1e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.748370] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.748547] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 896.749260] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a00b8a64-2218-4472-a941-b2ece28a8d39 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.754077] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 896.754077] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]520589fd-2b7b-94ca-b888-6217b6fc6e04" [ 896.754077] env[62183]: _type = "Task" [ 896.754077] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.763526] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]520589fd-2b7b-94ca-b888-6217b6fc6e04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.808766] env[62183]: DEBUG nova.compute.utils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.812954] env[62183]: DEBUG nova.compute.manager [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 896.813131] env[62183]: DEBUG nova.network.neutron [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.876404] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquiring lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.876733] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.876991] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquiring lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.877227] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.877440] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.879834] env[62183]: INFO nova.compute.manager [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Terminating instance [ 896.881820] env[62183]: DEBUG nova.compute.manager [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 896.882070] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 896.883087] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4a8591-85ab-4e56-84b2-ab3729c7f847 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.888955] env[62183]: DEBUG nova.policy [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88efcb2057e84413806f9ed809c260e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eda7e297c551407295b3c67fa0f1dbc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.899962] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.900271] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6b91a1f-e06d-41f9-9cc8-8441570355bf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.906535] env[62183]: DEBUG oslo_vmware.api [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 896.906535] env[62183]: value = "task-1387161" [ 896.906535] env[62183]: _type = "Task" [ 896.906535] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.914834] env[62183]: DEBUG oslo_vmware.api [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387161, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.039880] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34731507-453f-4826-ab80-5ec2497654c5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.048284] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97c25dc-f645-40ae-bce2-dd23f5d2410d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.082261] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a96c994-67a6-4c22-bcb5-f19698c93417 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.091861] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2511c585-f0d1-4819-b57e-0c35fc8e2740 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.112374] env[62183]: DEBUG nova.compute.provider_tree [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.114515] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.211828] env[62183]: DEBUG nova.compute.manager [req-ddc1bb55-ddc9-4ee3-a8de-a294931cfcbb req-bca37840-2c43-4c1a-a424-76759f7af597 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.212216] env[62183]: DEBUG nova.compute.manager [req-ddc1bb55-ddc9-4ee3-a8de-a294931cfcbb req-bca37840-2c43-4c1a-a424-76759f7af597 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing instance network info cache due to event network-changed-91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 897.212581] env[62183]: DEBUG oslo_concurrency.lockutils [req-ddc1bb55-ddc9-4ee3-a8de-a294931cfcbb req-bca37840-2c43-4c1a-a424-76759f7af597 service nova] Acquiring lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.212762] env[62183]: DEBUG oslo_concurrency.lockutils [req-ddc1bb55-ddc9-4ee3-a8de-a294931cfcbb req-bca37840-2c43-4c1a-a424-76759f7af597 service nova] Acquired lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.212934] env[62183]: DEBUG nova.network.neutron [req-ddc1bb55-ddc9-4ee3-a8de-a294931cfcbb req-bca37840-2c43-4c1a-a424-76759f7af597 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Refreshing network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.214506] env[62183]: DEBUG oslo_concurrency.lockutils [req-9252b6fd-2845-4817-947b-0f7a5759ab52 req-ac7220f2-b310-49b1-a6f8-6c2350ec7c9c service nova] Releasing lock "refresh_cache-14ea1b33-ffbd-4c48-84ea-443c10d8317d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.216554] env[62183]: INFO nova.compute.manager [-] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Took 1.30 seconds to deallocate network for instance. [ 897.267886] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]520589fd-2b7b-94ca-b888-6217b6fc6e04, 'name': SearchDatastore_Task, 'duration_secs': 0.008351} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.268910] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2e0f85e-9590-4754-aaea-300e0ecbb670 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.274655] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 897.274655] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5236c24e-75e8-5d5d-71d2-e9c9924bafde" [ 897.274655] env[62183]: _type = "Task" [ 897.274655] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.283304] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5236c24e-75e8-5d5d-71d2-e9c9924bafde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.313603] env[62183]: DEBUG nova.compute.manager [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.336346] env[62183]: DEBUG nova.network.neutron [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Successfully created port: 43abe649-25d9-41ef-9bd2-d09e3d07f261 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.423091] env[62183]: DEBUG oslo_vmware.api [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387161, 'name': PowerOffVM_Task, 'duration_secs': 0.253943} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.424415] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.424591] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.426729] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b53ed863-11e0-4783-8245-1c283c7eabe9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.499861] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.500092] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.500273] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Deleting the datastore file [datastore2] e3145bd7-85b2-4cc7-9d97-3e36a59b89cb {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.500542] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17cc954d-8a55-41d2-a35a-24b52f73bb13 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.507124] env[62183]: DEBUG oslo_vmware.api [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for the task: (returnval){ [ 897.507124] env[62183]: value = "task-1387163" [ 897.507124] env[62183]: _type = "Task" [ 897.507124] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.515755] env[62183]: DEBUG oslo_vmware.api [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387163, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.618443] env[62183]: DEBUG nova.scheduler.client.report [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 897.622343] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7096d50a-e6b0-4344-b0c0-01fdd240cf10 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "interface-ac555ffc-ce4e-4650-97fd-c26a3246fe4b-e7836918-44a9-4db8-8daf-35c2032e5cdb" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.868s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.722470] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.788810] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5236c24e-75e8-5d5d-71d2-e9c9924bafde, 'name': SearchDatastore_Task, 'duration_secs': 0.009975} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.789208] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.789572] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 14ea1b33-ffbd-4c48-84ea-443c10d8317d/14ea1b33-ffbd-4c48-84ea-443c10d8317d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.789928] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba438a4c-bf3d-4a9b-a75d-573b7607bc83 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.800459] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 897.800459] env[62183]: value = "task-1387164" [ 897.800459] env[62183]: _type = "Task" [ 897.800459] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.811324] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387164, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.839389] env[62183]: DEBUG nova.compute.manager [req-1862ea23-1c17-4513-902f-120659838f66 req-80dc30b6-e1d2-49f0-acdd-aecf94be9493 service nova] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Received event network-vif-deleted-d8838c27-a431-4260-88a0-4664366b6499 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.929989] env[62183]: DEBUG oslo_concurrency.lockutils [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "interface-1b0593f2-b712-4926-bfb8-30c92070ff6a-e7836918-44a9-4db8-8daf-35c2032e5cdb" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.930320] env[62183]: DEBUG oslo_concurrency.lockutils [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "interface-1b0593f2-b712-4926-bfb8-30c92070ff6a-e7836918-44a9-4db8-8daf-35c2032e5cdb" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.930684] env[62183]: DEBUG nova.objects.instance [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lazy-loading 'flavor' on Instance uuid 1b0593f2-b712-4926-bfb8-30c92070ff6a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.936193] env[62183]: DEBUG nova.network.neutron [req-ddc1bb55-ddc9-4ee3-a8de-a294931cfcbb req-bca37840-2c43-4c1a-a424-76759f7af597 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updated VIF entry in instance network info cache for port 91e1877d-c4f8-4ad7-8076-08c985c66e4c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.936193] env[62183]: DEBUG nova.network.neutron [req-ddc1bb55-ddc9-4ee3-a8de-a294931cfcbb req-bca37840-2c43-4c1a-a424-76759f7af597 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [{"id": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "address": "fa:16:3e:e4:8a:c9", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91e1877d-c4", "ovs_interfaceid": "91e1877d-c4f8-4ad7-8076-08c985c66e4c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.018383] env[62183]: DEBUG oslo_vmware.api [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Task: {'id': task-1387163, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14396} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.018666] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.018850] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.019081] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.019231] env[62183]: INFO nova.compute.manager [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Took 1.14 seconds to destroy the instance on the hypervisor. [ 898.019472] env[62183]: DEBUG oslo.service.loopingcall [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.019671] env[62183]: DEBUG nova.compute.manager [-] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 898.019767] env[62183]: DEBUG nova.network.neutron [-] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 898.314335] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387164, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.330060] env[62183]: DEBUG nova.compute.manager [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.362974] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.363256] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.363424] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.363614] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.363766] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.363919] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.364142] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.364303] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.364475] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.364689] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.364842] env[62183]: DEBUG nova.virt.hardware [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.365716] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eab2aa5-cd85-47b2-ba44-3ecb96d8f9e4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.373888] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d11457-6262-4855-bcea-1d9a597cea9c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.438761] env[62183]: DEBUG oslo_concurrency.lockutils [req-ddc1bb55-ddc9-4ee3-a8de-a294931cfcbb req-bca37840-2c43-4c1a-a424-76759f7af597 service nova] Releasing lock "refresh_cache-ac555ffc-ce4e-4650-97fd-c26a3246fe4b" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.594487] env[62183]: DEBUG nova.objects.instance [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lazy-loading 'pci_requests' on Instance uuid 1b0593f2-b712-4926-bfb8-30c92070ff6a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.630723] env[62183]: DEBUG oslo_concurrency.lockutils [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.325s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.634286] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.236s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.635198] env[62183]: INFO nova.compute.claims [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.811756] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387164, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552267} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.812040] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 14ea1b33-ffbd-4c48-84ea-443c10d8317d/14ea1b33-ffbd-4c48-84ea-443c10d8317d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.812268] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.812528] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-05a69e4f-1fd9-4edd-8984-a7133e1f920c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.819323] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 898.819323] env[62183]: value = "task-1387165" [ 898.819323] env[62183]: _type = "Task" [ 898.819323] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.827145] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387165, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.007600] env[62183]: DEBUG nova.network.neutron [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Successfully updated port: 43abe649-25d9-41ef-9bd2-d09e3d07f261 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.096873] env[62183]: DEBUG nova.objects.base [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Object Instance<1b0593f2-b712-4926-bfb8-30c92070ff6a> lazy-loaded attributes: flavor,pci_requests {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 899.097226] env[62183]: DEBUG nova.network.neutron [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 899.142488] env[62183]: DEBUG nova.network.neutron [-] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.166426] env[62183]: DEBUG nova.policy [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e0eee67c7954c80b4e1ea43b0abbcb5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebcc716f958942b588a6bfde78d2c00d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 899.199378] env[62183]: INFO nova.scheduler.client.report [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted allocation for migration 777f7ae0-1dd8-44e3-b8ed-0e0a79302012 [ 899.238363] env[62183]: DEBUG nova.compute.manager [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received event network-changed-999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.238599] env[62183]: DEBUG nova.compute.manager [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing instance network info cache due to event network-changed-999bf20b-d0e4-4229-a150-ec7c31e38cc6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 899.238832] env[62183]: DEBUG oslo_concurrency.lockutils [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] Acquiring lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.238983] env[62183]: DEBUG oslo_concurrency.lockutils [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] Acquired lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.239169] env[62183]: DEBUG nova.network.neutron [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing network info cache for port 999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 899.330440] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387165, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070234} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.330700] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.331487] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24ada9b-6bfd-49b5-888b-db3efc4b7135 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.352911] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 14ea1b33-ffbd-4c48-84ea-443c10d8317d/14ea1b33-ffbd-4c48-84ea-443c10d8317d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.353220] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e45aaab1-2d73-4c4c-84b0-62b37ed63f0a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.374068] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 899.374068] env[62183]: value = "task-1387166" [ 899.374068] env[62183]: _type = "Task" [ 899.374068] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.382334] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387166, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.510053] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "refresh_cache-4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.510223] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "refresh_cache-4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.510378] env[62183]: DEBUG nova.network.neutron [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.646930] env[62183]: INFO nova.compute.manager [-] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Took 1.63 seconds to deallocate network for instance. [ 899.704945] env[62183]: DEBUG oslo_concurrency.lockutils [None req-419b1288-f66d-412d-9ec8-54ce55666ab9 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.904s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.822180] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea61f8f6-8713-46cf-a63c-579c60859626 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.829974] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f71dff4-3a42-4518-b4a9-4ab4984109d9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.862196] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b1b3f8-0dae-477f-b342-6e51232555ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.870114] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3255dd-2938-4f95-9720-7808ea30211e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.887048] env[62183]: DEBUG nova.compute.provider_tree [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.893786] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387166, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.919956] env[62183]: DEBUG nova.compute.manager [req-40fc5dd8-3c06-4611-9c38-3337d0c1653b req-64112ea2-6183-4364-90bc-f94380f29180 service nova] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Received event network-vif-deleted-d3884443-e294-4422-9ebe-924202ddaeab {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.992151] env[62183]: DEBUG nova.network.neutron [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updated VIF entry in instance network info cache for port 999bf20b-d0e4-4229-a150-ec7c31e38cc6. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 899.992536] env[62183]: DEBUG nova.network.neutron [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updating instance_info_cache with network_info: [{"id": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "address": "fa:16:3e:46:24:80", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap999bf20b-d0", "ovs_interfaceid": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.058128] env[62183]: DEBUG nova.network.neutron [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.158683] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.228007] env[62183]: DEBUG nova.network.neutron [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Updating instance_info_cache with network_info: [{"id": "43abe649-25d9-41ef-9bd2-d09e3d07f261", "address": "fa:16:3e:b0:00:b1", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43abe649-25", "ovs_interfaceid": "43abe649-25d9-41ef-9bd2-d09e3d07f261", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.388143] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387166, 'name': ReconfigVM_Task, 'duration_secs': 0.998031} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.388143] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 14ea1b33-ffbd-4c48-84ea-443c10d8317d/14ea1b33-ffbd-4c48-84ea-443c10d8317d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.389013] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd3a7267-3cbf-4306-a69c-29888477d937 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.395993] env[62183]: DEBUG nova.scheduler.client.report [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 900.400452] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 900.400452] env[62183]: value = "task-1387167" [ 900.400452] env[62183]: _type = "Task" [ 900.400452] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.411858] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387167, 'name': Rename_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.495397] env[62183]: DEBUG oslo_concurrency.lockutils [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] Releasing lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.495698] env[62183]: DEBUG nova.compute.manager [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Received event network-vif-plugged-43abe649-25d9-41ef-9bd2-d09e3d07f261 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.495900] env[62183]: DEBUG oslo_concurrency.lockutils [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] Acquiring lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.496142] env[62183]: DEBUG oslo_concurrency.lockutils [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] Lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.496312] env[62183]: DEBUG oslo_concurrency.lockutils [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] Lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.496483] env[62183]: DEBUG nova.compute.manager [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] No waiting events found dispatching network-vif-plugged-43abe649-25d9-41ef-9bd2-d09e3d07f261 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 900.496653] env[62183]: WARNING nova.compute.manager [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Received unexpected event network-vif-plugged-43abe649-25d9-41ef-9bd2-d09e3d07f261 for instance with vm_state building and task_state spawning. [ 900.496817] env[62183]: DEBUG nova.compute.manager [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Received event network-changed-43abe649-25d9-41ef-9bd2-d09e3d07f261 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.497068] env[62183]: DEBUG nova.compute.manager [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Refreshing instance network info cache due to event network-changed-43abe649-25d9-41ef-9bd2-d09e3d07f261. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.497268] env[62183]: DEBUG oslo_concurrency.lockutils [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] Acquiring lock "refresh_cache-4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.731144] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "refresh_cache-4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.731504] env[62183]: DEBUG nova.compute.manager [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Instance network_info: |[{"id": "43abe649-25d9-41ef-9bd2-d09e3d07f261", "address": "fa:16:3e:b0:00:b1", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43abe649-25", "ovs_interfaceid": "43abe649-25d9-41ef-9bd2-d09e3d07f261", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 900.731851] env[62183]: DEBUG oslo_concurrency.lockutils [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] Acquired lock "refresh_cache-4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.732088] env[62183]: DEBUG nova.network.neutron [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Refreshing network info cache for port 43abe649-25d9-41ef-9bd2-d09e3d07f261 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.733423] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:00:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43abe649-25d9-41ef-9bd2-d09e3d07f261', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.741683] env[62183]: DEBUG oslo.service.loopingcall [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.742624] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.742878] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-baa6ffa7-d561-4659-9887-914eb467f763 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.762633] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.762633] env[62183]: value = "task-1387168" [ 900.762633] env[62183]: _type = "Task" [ 900.762633] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.772112] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387168, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.902427] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.903172] env[62183]: DEBUG nova.compute.manager [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 900.906529] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.184s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.906772] env[62183]: DEBUG nova.objects.instance [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lazy-loading 'resources' on Instance uuid 850d40d8-565a-49a2-a27f-3de2a8dc7e30 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.916813] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387167, 'name': Rename_Task, 'duration_secs': 0.437752} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.917701] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 900.918016] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2b1b107-80c9-4dbe-8c32-c83705cb79ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.924172] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 900.924172] env[62183]: value = "task-1387169" [ 900.924172] env[62183]: _type = "Task" [ 900.924172] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.932528] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387169, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.156642] env[62183]: DEBUG nova.network.neutron [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Successfully updated port: e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.271136] env[62183]: DEBUG nova.compute.manager [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received event network-vif-plugged-e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.271417] env[62183]: DEBUG oslo_concurrency.lockutils [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] Acquiring lock "1b0593f2-b712-4926-bfb8-30c92070ff6a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.271576] env[62183]: DEBUG oslo_concurrency.lockutils [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.271745] env[62183]: DEBUG oslo_concurrency.lockutils [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.272063] env[62183]: DEBUG nova.compute.manager [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] No waiting events found dispatching network-vif-plugged-e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 901.272134] env[62183]: WARNING nova.compute.manager [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received unexpected event network-vif-plugged-e7836918-44a9-4db8-8daf-35c2032e5cdb for instance with vm_state active and task_state None. [ 901.272256] env[62183]: DEBUG nova.compute.manager [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received event network-changed-e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.272403] env[62183]: DEBUG nova.compute.manager [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing instance network info cache due to event network-changed-e7836918-44a9-4db8-8daf-35c2032e5cdb. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 901.272623] env[62183]: DEBUG oslo_concurrency.lockutils [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] Acquiring lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.273024] env[62183]: DEBUG oslo_concurrency.lockutils [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] Acquired lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.273248] env[62183]: DEBUG nova.network.neutron [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Refreshing network info cache for port e7836918-44a9-4db8-8daf-35c2032e5cdb {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.278465] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387168, 'name': CreateVM_Task, 'duration_secs': 0.313758} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.278824] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.279489] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.279671] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.280018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.280545] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb06ca27-6aed-449d-b831-0d0ea2689d49 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.288408] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 901.288408] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52c1c0fc-a7cb-3d8d-05be-a6cf497eb3fc" [ 901.288408] env[62183]: _type = "Task" [ 901.288408] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.297514] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c1c0fc-a7cb-3d8d-05be-a6cf497eb3fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.415111] env[62183]: DEBUG nova.compute.utils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.416676] env[62183]: DEBUG nova.compute.manager [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.416840] env[62183]: DEBUG nova.network.neutron [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 901.435145] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387169, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.457113] env[62183]: DEBUG nova.policy [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7deeb5ccaca42e29f0cb145a17e1097', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37d4936e2dc34e9d9ed73455081ee4ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 901.571444] env[62183]: DEBUG nova.network.neutron [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Updated VIF entry in instance network info cache for port 43abe649-25d9-41ef-9bd2-d09e3d07f261. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.571767] env[62183]: DEBUG nova.network.neutron [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Updating instance_info_cache with network_info: [{"id": "43abe649-25d9-41ef-9bd2-d09e3d07f261", "address": "fa:16:3e:b0:00:b1", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43abe649-25", "ovs_interfaceid": "43abe649-25d9-41ef-9bd2-d09e3d07f261", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.596272] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d193763f-40e5-43e1-b737-c1de1312ae7e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.604314] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484fabb4-9513-4d1d-b6fd-79a6b74a72fe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.635380] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8c6de2-8975-4dc2-a711-b76ec105dd2e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.642655] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435a7696-0fa1-46cf-b56a-9c77b1ce629b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.655832] env[62183]: DEBUG nova.compute.provider_tree [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.662507] env[62183]: DEBUG oslo_concurrency.lockutils [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.706838] env[62183]: DEBUG nova.network.neutron [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Successfully created port: 34118d05-2bab-4122-82fa-0a5ae337cbbe {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 901.750634] env[62183]: DEBUG oslo_concurrency.lockutils [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.750634] env[62183]: DEBUG oslo_concurrency.lockutils [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.750730] env[62183]: DEBUG oslo_concurrency.lockutils [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.750875] env[62183]: DEBUG oslo_concurrency.lockutils [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.751072] env[62183]: DEBUG oslo_concurrency.lockutils [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.753455] env[62183]: INFO nova.compute.manager [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Terminating instance [ 901.755871] env[62183]: DEBUG nova.compute.manager [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.755871] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 901.756530] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2393e7-04e9-4e19-ad11-ebc05a0249f3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.764439] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.764672] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73bec845-865d-4ff6-ba79-427edbbcb286 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.771377] env[62183]: DEBUG oslo_vmware.api [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 901.771377] env[62183]: value = "task-1387170" [ 901.771377] env[62183]: _type = "Task" [ 901.771377] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.781448] env[62183]: DEBUG oslo_vmware.api [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387170, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.800259] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c1c0fc-a7cb-3d8d-05be-a6cf497eb3fc, 'name': SearchDatastore_Task, 'duration_secs': 0.009861} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.800880] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.801163] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.801420] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.801573] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.801940] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.802618] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2f86ed2-02d1-4676-8d5c-b9e51f76394c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.814254] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.814254] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.814254] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4e61f04-a085-4380-bbd3-2457c6285489 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.826048] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 901.826048] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52be02c9-f347-bb25-ca35-2c943ba759b5" [ 901.826048] env[62183]: _type = "Task" [ 901.826048] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.837216] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52be02c9-f347-bb25-ca35-2c943ba759b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.920364] env[62183]: DEBUG nova.compute.manager [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 901.936013] env[62183]: DEBUG oslo_vmware.api [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387169, 'name': PowerOnVM_Task, 'duration_secs': 0.979959} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.936419] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 901.936632] env[62183]: INFO nova.compute.manager [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Took 8.61 seconds to spawn the instance on the hypervisor. [ 901.936815] env[62183]: DEBUG nova.compute.manager [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.941027] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ddfd1b-645c-43c1-9f77-6be313eb9be4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.076129] env[62183]: DEBUG oslo_concurrency.lockutils [req-19f33f12-6ebe-43fa-94b7-6935f8b7d1ca req-e3f1d6d6-d7a7-46d3-9ef7-d4296239fa47 service nova] Releasing lock "refresh_cache-4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.095525] env[62183]: DEBUG nova.network.neutron [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Added VIF to instance network info cache for port e7836918-44a9-4db8-8daf-35c2032e5cdb. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 902.096259] env[62183]: DEBUG nova.network.neutron [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updating instance_info_cache with network_info: [{"id": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "address": "fa:16:3e:46:24:80", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap999bf20b-d0", "ovs_interfaceid": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e7836918-44a9-4db8-8daf-35c2032e5cdb", "address": "fa:16:3e:c8:56:ec", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7836918-44", "ovs_interfaceid": "e7836918-44a9-4db8-8daf-35c2032e5cdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.161570] env[62183]: DEBUG nova.scheduler.client.report [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 902.282138] env[62183]: DEBUG oslo_vmware.api [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387170, 'name': PowerOffVM_Task, 'duration_secs': 0.250907} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.282512] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.282804] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.283167] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cfdc1b66-0329-47d5-884d-db337b280652 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.291125] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "28d279bd-206e-4ba0-bf84-2980ab41f38d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.291125] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.336659] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52be02c9-f347-bb25-ca35-2c943ba759b5, 'name': SearchDatastore_Task, 'duration_secs': 0.011612} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.336891] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9068e4c0-3916-48f8-b2b3-bbbe73716368 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.342894] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 902.342894] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52043fe5-6650-59a0-ec5c-78f01aa82751" [ 902.342894] env[62183]: _type = "Task" [ 902.342894] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.350886] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52043fe5-6650-59a0-ec5c-78f01aa82751, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.357674] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 902.357862] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 902.358063] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleting the datastore file [datastore1] 503786ca-dba4-43c1-9a25-9f1cbac9a6a4 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.358329] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d579052e-f103-4a8f-9db3-5ab7484b19cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.365368] env[62183]: DEBUG oslo_vmware.api [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 902.365368] env[62183]: value = "task-1387172" [ 902.365368] env[62183]: _type = "Task" [ 902.365368] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.373615] env[62183]: DEBUG oslo_vmware.api [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.462492] env[62183]: INFO nova.compute.manager [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Took 14.07 seconds to build instance. [ 902.599062] env[62183]: DEBUG oslo_concurrency.lockutils [req-fcbb535a-cb6a-4d0e-972e-76bf7aa360cd req-e432e3d3-503a-4709-86e8-d7e7bb182013 service nova] Releasing lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.599603] env[62183]: DEBUG oslo_concurrency.lockutils [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.599868] env[62183]: DEBUG nova.network.neutron [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.666641] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.669195] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.511s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.669464] env[62183]: DEBUG nova.objects.instance [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lazy-loading 'resources' on Instance uuid e3145bd7-85b2-4cc7-9d97-3e36a59b89cb {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.688562] env[62183]: INFO nova.scheduler.client.report [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleted allocations for instance 850d40d8-565a-49a2-a27f-3de2a8dc7e30 [ 902.793429] env[62183]: DEBUG nova.compute.manager [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.853755] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52043fe5-6650-59a0-ec5c-78f01aa82751, 'name': SearchDatastore_Task, 'duration_secs': 0.015264} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.854106] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.854424] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7/4de35fd3-7ef8-4a0b-b180-8b36e308e2e7.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 902.854701] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1914587-8b9f-4387-953f-3bfdfeeff02e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.862041] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 902.862041] env[62183]: value = "task-1387173" [ 902.862041] env[62183]: _type = "Task" [ 902.862041] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.872326] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387173, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.879117] env[62183]: DEBUG oslo_vmware.api [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198585} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.879594] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.879771] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 902.880329] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 902.880329] env[62183]: INFO nova.compute.manager [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Took 1.12 seconds to destroy the instance on the hypervisor. [ 902.880498] env[62183]: DEBUG oslo.service.loopingcall [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.880566] env[62183]: DEBUG nova.compute.manager [-] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 902.880654] env[62183]: DEBUG nova.network.neutron [-] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 902.937941] env[62183]: DEBUG nova.compute.manager [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 902.960265] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 902.960552] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 902.960729] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.960954] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 902.961158] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.961332] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 902.961566] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 902.961747] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 902.961951] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 902.962151] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 902.962333] env[62183]: DEBUG nova.virt.hardware [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.962854] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bff3cc6c-05f8-4e14-9563-f42ecd621786 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.585s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.963653] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45cd3052-0b97-450d-9151-fc239f688015 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.972055] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db567f43-5b4b-4544-8430-cef12b754c49 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.148645] env[62183]: WARNING nova.network.neutron [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] ece15cf2-6921-4b76-9d7e-f313bcaa2f48 already exists in list: networks containing: ['ece15cf2-6921-4b76-9d7e-f313bcaa2f48']. ignoring it [ 903.148869] env[62183]: WARNING nova.network.neutron [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] ece15cf2-6921-4b76-9d7e-f313bcaa2f48 already exists in list: networks containing: ['ece15cf2-6921-4b76-9d7e-f313bcaa2f48']. ignoring it [ 903.149055] env[62183]: WARNING nova.network.neutron [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] e7836918-44a9-4db8-8daf-35c2032e5cdb already exists in list: port_ids containing: ['e7836918-44a9-4db8-8daf-35c2032e5cdb']. ignoring it [ 903.195424] env[62183]: DEBUG nova.compute.manager [req-fe67d3b6-57f7-4fe8-add9-ec0360e5d663 req-42186e54-55c4-49a5-995c-b16531395fe7 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Received event network-vif-plugged-34118d05-2bab-4122-82fa-0a5ae337cbbe {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.195698] env[62183]: DEBUG oslo_concurrency.lockutils [req-fe67d3b6-57f7-4fe8-add9-ec0360e5d663 req-42186e54-55c4-49a5-995c-b16531395fe7 service nova] Acquiring lock "46aa1062-3fcd-4581-8647-32e1372e9491-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.195883] env[62183]: DEBUG oslo_concurrency.lockutils [req-fe67d3b6-57f7-4fe8-add9-ec0360e5d663 req-42186e54-55c4-49a5-995c-b16531395fe7 service nova] Lock "46aa1062-3fcd-4581-8647-32e1372e9491-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.196070] env[62183]: DEBUG oslo_concurrency.lockutils [req-fe67d3b6-57f7-4fe8-add9-ec0360e5d663 req-42186e54-55c4-49a5-995c-b16531395fe7 service nova] Lock "46aa1062-3fcd-4581-8647-32e1372e9491-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.196872] env[62183]: DEBUG nova.compute.manager [req-fe67d3b6-57f7-4fe8-add9-ec0360e5d663 req-42186e54-55c4-49a5-995c-b16531395fe7 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] No waiting events found dispatching network-vif-plugged-34118d05-2bab-4122-82fa-0a5ae337cbbe {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 903.197131] env[62183]: WARNING nova.compute.manager [req-fe67d3b6-57f7-4fe8-add9-ec0360e5d663 req-42186e54-55c4-49a5-995c-b16531395fe7 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Received unexpected event network-vif-plugged-34118d05-2bab-4122-82fa-0a5ae337cbbe for instance with vm_state building and task_state spawning. [ 903.201941] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d256a2a2-4348-46f7-9631-b3a5f1a2bf66 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "850d40d8-565a-49a2-a27f-3de2a8dc7e30" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.405s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.323668] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.371374] env[62183]: DEBUG nova.network.neutron [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Successfully updated port: 34118d05-2bab-4122-82fa-0a5ae337cbbe {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 903.376191] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387173, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.381564] env[62183]: DEBUG nova.compute.manager [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.382510] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0427971c-afef-4d3b-a05f-18b361696d2e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.408952] env[62183]: DEBUG nova.compute.manager [req-2ae44a66-60fd-4537-ae31-1bbc31588c62 req-1a4a0398-7e20-48f3-9a52-3c63128109e2 service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Received event network-vif-deleted-d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.409088] env[62183]: INFO nova.compute.manager [req-2ae44a66-60fd-4537-ae31-1bbc31588c62 req-1a4a0398-7e20-48f3-9a52-3c63128109e2 service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Neutron deleted interface d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29; detaching it from the instance and deleting it from the info cache [ 903.409257] env[62183]: DEBUG nova.network.neutron [req-2ae44a66-60fd-4537-ae31-1bbc31588c62 req-1a4a0398-7e20-48f3-9a52-3c63128109e2 service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.415625] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75624a10-dfaf-457a-a6df-e66fe8e69f52 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.424156] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92572991-217e-4a1d-b101-795c4847b189 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.465674] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d306baf1-5ea7-435f-bd99-ad6f84edcfae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.474850] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cfa4b4-fa5b-45c6-91a1-f9a8619cebea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.488937] env[62183]: DEBUG nova.compute.provider_tree [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.611416] env[62183]: DEBUG nova.network.neutron [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updating instance_info_cache with network_info: [{"id": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "address": "fa:16:3e:46:24:80", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap999bf20b-d0", "ovs_interfaceid": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e7836918-44a9-4db8-8daf-35c2032e5cdb", "address": "fa:16:3e:c8:56:ec", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7836918-44", "ovs_interfaceid": "e7836918-44a9-4db8-8daf-35c2032e5cdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.698231] env[62183]: DEBUG nova.network.neutron [-] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.817759] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b6186a-74b3-410b-b449-9df6c23c2350 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.824260] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d2404ead-744c-46b6-b3a2-70d13cf63c1f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Suspending the VM {{(pid=62183) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 903.824487] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-0607777d-4f3e-4c64-b07a-ddaa92eb32cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.829910] env[62183]: DEBUG oslo_vmware.api [None req-d2404ead-744c-46b6-b3a2-70d13cf63c1f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 903.829910] env[62183]: value = "task-1387174" [ 903.829910] env[62183]: _type = "Task" [ 903.829910] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.837473] env[62183]: DEBUG oslo_vmware.api [None req-d2404ead-744c-46b6-b3a2-70d13cf63c1f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387174, 'name': SuspendVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.872977] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "refresh_cache-46aa1062-3fcd-4581-8647-32e1372e9491" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.873180] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "refresh_cache-46aa1062-3fcd-4581-8647-32e1372e9491" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.873404] env[62183]: DEBUG nova.network.neutron [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 903.874608] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387173, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645081} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.875362] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7/4de35fd3-7ef8-4a0b-b180-8b36e308e2e7.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.875362] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.875646] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc6466e5-ea48-4c87-b710-faf0c50d0734 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.881774] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 903.881774] env[62183]: value = "task-1387175" [ 903.881774] env[62183]: _type = "Task" [ 903.881774] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.890361] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387175, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.896743] env[62183]: INFO nova.compute.manager [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] instance snapshotting [ 903.897379] env[62183]: DEBUG nova.objects.instance [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'flavor' on Instance uuid a948464b-63aa-4bc8-9885-228049e96d37 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.912605] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c7b2419-0817-43be-97cf-20c707494a7f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.921933] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa678569-cf09-453e-85ed-80d4ef47e7c9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.956448] env[62183]: DEBUG nova.compute.manager [req-2ae44a66-60fd-4537-ae31-1bbc31588c62 req-1a4a0398-7e20-48f3-9a52-3c63128109e2 service nova] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Detach interface failed, port_id=d2f9a654-1fa3-48a5-a9ba-c6de69dc5f29, reason: Instance 503786ca-dba4-43c1-9a25-9f1cbac9a6a4 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 903.992882] env[62183]: DEBUG nova.scheduler.client.report [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 904.114092] env[62183]: DEBUG oslo_concurrency.lockutils [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.114891] env[62183]: DEBUG oslo_concurrency.lockutils [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.115115] env[62183]: DEBUG oslo_concurrency.lockutils [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.116081] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02732d1-7838-4c73-8950-1ab37e796d58 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.134061] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.134334] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.134468] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.134664] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.134861] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.135065] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.135257] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.135421] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.135591] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.135779] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.135998] env[62183]: DEBUG nova.virt.hardware [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.142825] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Reconfiguring VM to attach interface {{(pid=62183) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 904.143198] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94a15771-086b-4b98-ad90-885b603ab732 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.161348] env[62183]: DEBUG oslo_vmware.api [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 904.161348] env[62183]: value = "task-1387176" [ 904.161348] env[62183]: _type = "Task" [ 904.161348] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.170011] env[62183]: DEBUG oslo_vmware.api [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387176, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.200819] env[62183]: INFO nova.compute.manager [-] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Took 1.32 seconds to deallocate network for instance. [ 904.340694] env[62183]: DEBUG oslo_vmware.api [None req-d2404ead-744c-46b6-b3a2-70d13cf63c1f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387174, 'name': SuspendVM_Task} progress is 54%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.391940] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387175, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089985} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.393228] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.394296] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c18f813-df85-4a9a-8dfd-d31d224d03ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.416918] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7/4de35fd3-7ef8-4a0b-b180-8b36e308e2e7.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.417836] env[62183]: DEBUG nova.network.neutron [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 904.419905] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf487dc9-090a-4b0d-bc1f-7cbd8f6c8c92 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.434245] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb99542-7078-4c62-a657-7a5772c2a90a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.456383] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ac7e9a-b32a-4cb4-9eac-d4eb4598b3eb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.458944] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 904.458944] env[62183]: value = "task-1387177" [ 904.458944] env[62183]: _type = "Task" [ 904.458944] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.471014] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387177, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.499083] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.830s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.501372] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.178s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.502814] env[62183]: INFO nova.compute.claims [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.523502] env[62183]: INFO nova.scheduler.client.report [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Deleted allocations for instance e3145bd7-85b2-4cc7-9d97-3e36a59b89cb [ 904.591065] env[62183]: DEBUG nova.network.neutron [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Updating instance_info_cache with network_info: [{"id": "34118d05-2bab-4122-82fa-0a5ae337cbbe", "address": "fa:16:3e:a5:b4:38", "network": {"id": "bef4069c-60bb-4ca3-b068-b7d33e5c9d90", "bridge": "br-int", "label": "tempest-ImagesTestJSON-798269247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37d4936e2dc34e9d9ed73455081ee4ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34118d05-2b", "ovs_interfaceid": "34118d05-2bab-4122-82fa-0a5ae337cbbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.673283] env[62183]: DEBUG oslo_vmware.api [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387176, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.707762] env[62183]: DEBUG oslo_concurrency.lockutils [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.840936] env[62183]: DEBUG oslo_vmware.api [None req-d2404ead-744c-46b6-b3a2-70d13cf63c1f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387174, 'name': SuspendVM_Task, 'duration_secs': 0.869095} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.841229] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d2404ead-744c-46b6-b3a2-70d13cf63c1f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Suspended the VM {{(pid=62183) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 904.841426] env[62183]: DEBUG nova.compute.manager [None req-d2404ead-744c-46b6-b3a2-70d13cf63c1f tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.842209] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9944b43-eefe-4491-a42f-8a76c7b97952 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.970255] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Creating Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 904.970586] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387177, 'name': ReconfigVM_Task, 'duration_secs': 0.272379} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.970809] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-46b0b519-8e2f-4a69-a20f-b67c31c1f349 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.972841] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7/4de35fd3-7ef8-4a0b-b180-8b36e308e2e7.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.973426] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38b2b665-4eba-43f3-9c94-dff6cc655753 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.979128] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 904.979128] env[62183]: value = "task-1387178" [ 904.979128] env[62183]: _type = "Task" [ 904.979128] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.982707] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 904.982707] env[62183]: value = "task-1387179" [ 904.982707] env[62183]: _type = "Task" [ 904.982707] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.988459] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387178, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.993037] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387179, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.032759] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f272f305-6551-4696-b86f-eab4cb953dad tempest-ServersTestManualDisk-625877190 tempest-ServersTestManualDisk-625877190-project-member] Lock "e3145bd7-85b2-4cc7-9d97-3e36a59b89cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.156s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.094757] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "refresh_cache-46aa1062-3fcd-4581-8647-32e1372e9491" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.095157] env[62183]: DEBUG nova.compute.manager [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Instance network_info: |[{"id": "34118d05-2bab-4122-82fa-0a5ae337cbbe", "address": "fa:16:3e:a5:b4:38", "network": {"id": "bef4069c-60bb-4ca3-b068-b7d33e5c9d90", "bridge": "br-int", "label": "tempest-ImagesTestJSON-798269247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37d4936e2dc34e9d9ed73455081ee4ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34118d05-2b", "ovs_interfaceid": "34118d05-2bab-4122-82fa-0a5ae337cbbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.095616] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:b4:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '11032cc2-b275-48d2-9c40-9455ea7d49e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34118d05-2bab-4122-82fa-0a5ae337cbbe', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.103225] env[62183]: DEBUG oslo.service.loopingcall [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.103443] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.103668] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25420fbe-64ac-4cba-b6d6-e8900f7c79a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.122776] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.122776] env[62183]: value = "task-1387180" [ 905.122776] env[62183]: _type = "Task" [ 905.122776] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.130086] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387180, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.173672] env[62183]: DEBUG oslo_vmware.api [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387176, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.266907] env[62183]: DEBUG nova.compute.manager [req-e0e404d6-c32b-4548-9755-e15de330bf4a req-38781f84-0fea-423f-91e0-e54e286610c4 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Received event network-changed-34118d05-2bab-4122-82fa-0a5ae337cbbe {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.267182] env[62183]: DEBUG nova.compute.manager [req-e0e404d6-c32b-4548-9755-e15de330bf4a req-38781f84-0fea-423f-91e0-e54e286610c4 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Refreshing instance network info cache due to event network-changed-34118d05-2bab-4122-82fa-0a5ae337cbbe. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.267857] env[62183]: DEBUG oslo_concurrency.lockutils [req-e0e404d6-c32b-4548-9755-e15de330bf4a req-38781f84-0fea-423f-91e0-e54e286610c4 service nova] Acquiring lock "refresh_cache-46aa1062-3fcd-4581-8647-32e1372e9491" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.267857] env[62183]: DEBUG oslo_concurrency.lockutils [req-e0e404d6-c32b-4548-9755-e15de330bf4a req-38781f84-0fea-423f-91e0-e54e286610c4 service nova] Acquired lock "refresh_cache-46aa1062-3fcd-4581-8647-32e1372e9491" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.267857] env[62183]: DEBUG nova.network.neutron [req-e0e404d6-c32b-4548-9755-e15de330bf4a req-38781f84-0fea-423f-91e0-e54e286610c4 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Refreshing network info cache for port 34118d05-2bab-4122-82fa-0a5ae337cbbe {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 905.492612] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387179, 'name': Rename_Task, 'duration_secs': 0.137412} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.495684] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.496264] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387178, 'name': CreateSnapshot_Task, 'duration_secs': 0.497712} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.496465] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d7623b9-f9e3-4403-8cc9-129d82209540 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.497868] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Created Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 905.498564] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c335eb1c-095a-421d-b120-7b5f92a04edb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.508444] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 905.508444] env[62183]: value = "task-1387181" [ 905.508444] env[62183]: _type = "Task" [ 905.508444] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.635187] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387180, 'name': CreateVM_Task, 'duration_secs': 0.322507} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.635358] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.636051] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.636238] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.636551] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.636808] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47423a72-7b08-43b7-a597-8a221aeac03c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.641264] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 905.641264] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]527e7e9c-7150-490d-0509-9e77c597742e" [ 905.641264] env[62183]: _type = "Task" [ 905.641264] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.652216] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527e7e9c-7150-490d-0509-9e77c597742e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.675494] env[62183]: DEBUG oslo_vmware.api [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387176, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.682095] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13eb0b30-0f3d-424f-acdc-606f81615c5e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.690212] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053367da-882f-4601-8565-15cbc8657634 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.721568] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d756f4aa-d462-4f98-a0b7-410f9202b6b6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.729978] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2019d1b8-96e8-4eab-a89a-2514a6d9d769 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.744758] env[62183]: DEBUG nova.compute.provider_tree [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.994899] env[62183]: DEBUG nova.network.neutron [req-e0e404d6-c32b-4548-9755-e15de330bf4a req-38781f84-0fea-423f-91e0-e54e286610c4 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Updated VIF entry in instance network info cache for port 34118d05-2bab-4122-82fa-0a5ae337cbbe. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 905.995401] env[62183]: DEBUG nova.network.neutron [req-e0e404d6-c32b-4548-9755-e15de330bf4a req-38781f84-0fea-423f-91e0-e54e286610c4 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Updating instance_info_cache with network_info: [{"id": "34118d05-2bab-4122-82fa-0a5ae337cbbe", "address": "fa:16:3e:a5:b4:38", "network": {"id": "bef4069c-60bb-4ca3-b068-b7d33e5c9d90", "bridge": "br-int", "label": "tempest-ImagesTestJSON-798269247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37d4936e2dc34e9d9ed73455081ee4ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "11032cc2-b275-48d2-9c40-9455ea7d49e3", "external-id": "nsx-vlan-transportzone-226", "segmentation_id": 226, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34118d05-2b", "ovs_interfaceid": "34118d05-2bab-4122-82fa-0a5ae337cbbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.017947] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Creating linked-clone VM from snapshot {{(pid=62183) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 906.021720] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0ddc60be-3574-4801-894f-e6d3a87eea8a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.038412] env[62183]: DEBUG oslo_vmware.api [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387181, 'name': PowerOnVM_Task, 'duration_secs': 0.528005} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.040360] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 906.040654] env[62183]: INFO nova.compute.manager [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Took 7.71 seconds to spawn the instance on the hypervisor. [ 906.040852] env[62183]: DEBUG nova.compute.manager [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.041521] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 906.041521] env[62183]: value = "task-1387182" [ 906.041521] env[62183]: _type = "Task" [ 906.041521] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.042441] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7787672d-b698-419d-a52f-ef2a74951841 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.060034] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387182, 'name': CloneVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.155500] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527e7e9c-7150-490d-0509-9e77c597742e, 'name': SearchDatastore_Task, 'duration_secs': 0.012535} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.155957] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.156289] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.156577] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.156843] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.157154] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.157579] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1847402-4464-44d2-b454-3786ff2bc827 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.170744] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.170928] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.172014] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0220583e-e7d8-47fc-b7d9-c8e890e610f3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.177316] env[62183]: DEBUG oslo_vmware.api [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387176, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.180266] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 906.180266] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a612cf-af67-0a25-b0f8-5a5928cc4212" [ 906.180266] env[62183]: _type = "Task" [ 906.180266] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.187813] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a612cf-af67-0a25-b0f8-5a5928cc4212, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.248661] env[62183]: DEBUG nova.scheduler.client.report [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 906.499105] env[62183]: DEBUG oslo_concurrency.lockutils [req-e0e404d6-c32b-4548-9755-e15de330bf4a req-38781f84-0fea-423f-91e0-e54e286610c4 service nova] Releasing lock "refresh_cache-46aa1062-3fcd-4581-8647-32e1372e9491" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.524251] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.524251] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.524251] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.524251] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.524251] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.525599] env[62183]: INFO nova.compute.manager [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Terminating instance [ 906.527953] env[62183]: DEBUG nova.compute.manager [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 906.528368] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.529339] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee4549d-2c66-4296-92a0-892efc3120b9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.539035] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.539035] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7411ace3-38f4-452d-827b-c7998b5c7a64 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.553977] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387182, 'name': CloneVM_Task} progress is 93%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.566714] env[62183]: INFO nova.compute.manager [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Took 12.55 seconds to build instance. [ 906.629846] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.629846] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.629846] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleting the datastore file [datastore1] 14ea1b33-ffbd-4c48-84ea-443c10d8317d {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.629846] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b20009d1-625d-4dc8-a65c-b9f6d76d860a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.635589] env[62183]: DEBUG oslo_vmware.api [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 906.635589] env[62183]: value = "task-1387184" [ 906.635589] env[62183]: _type = "Task" [ 906.635589] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.645408] env[62183]: DEBUG oslo_vmware.api [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387184, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.674989] env[62183]: DEBUG oslo_vmware.api [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387176, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.689814] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a612cf-af67-0a25-b0f8-5a5928cc4212, 'name': SearchDatastore_Task, 'duration_secs': 0.013702} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.693189] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d347860-dfa3-4300-b795-f39fb7e02068 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.695854] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 906.695854] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52bcb98d-8b32-d546-8cef-3f36414b2181" [ 906.695854] env[62183]: _type = "Task" [ 906.695854] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.703868] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bcb98d-8b32-d546-8cef-3f36414b2181, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.757339] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.757339] env[62183]: DEBUG nova.compute.manager [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 906.758665] env[62183]: DEBUG oslo_concurrency.lockutils [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.051s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.761669] env[62183]: DEBUG oslo_concurrency.lockutils [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.790674] env[62183]: INFO nova.scheduler.client.report [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted allocations for instance 503786ca-dba4-43c1-9a25-9f1cbac9a6a4 [ 907.058398] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387182, 'name': CloneVM_Task} progress is 93%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.071247] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4e8d912-d2d2-4b08-94ca-21cd9e48ace4 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.068s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.150831] env[62183]: DEBUG oslo_vmware.api [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387184, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.177394] env[62183]: DEBUG oslo_vmware.api [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387176, 'name': ReconfigVM_Task, 'duration_secs': 2.918784} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.177930] env[62183]: DEBUG oslo_concurrency.lockutils [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.178172] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Reconfigured VM to attach interface {{(pid=62183) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 907.211075] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bcb98d-8b32-d546-8cef-3f36414b2181, 'name': SearchDatastore_Task, 'duration_secs': 0.025126} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.211520] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.212369] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 46aa1062-3fcd-4581-8647-32e1372e9491/46aa1062-3fcd-4581-8647-32e1372e9491.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.213026] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c668d696-bdbb-4025-8e7a-b74cb78cf218 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.222061] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 907.222061] env[62183]: value = "task-1387185" [ 907.222061] env[62183]: _type = "Task" [ 907.222061] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.230419] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387185, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.264048] env[62183]: DEBUG nova.compute.utils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.265645] env[62183]: DEBUG nova.compute.manager [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 907.265868] env[62183]: DEBUG nova.network.neutron [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 907.303829] env[62183]: DEBUG oslo_concurrency.lockutils [None req-938466f6-f1b2-47e4-a4fe-8b09c3ec043a tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "503786ca-dba4-43c1-9a25-9f1cbac9a6a4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.553s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.323184] env[62183]: DEBUG nova.policy [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89ad04efd23b40e6a3eea5b3e6fd021e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b61c6a99cbda435481bb72f20929f03f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 907.563462] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387182, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.653359] env[62183]: DEBUG oslo_vmware.api [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387184, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.567444} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.656561] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.656749] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 907.659397] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 907.659397] env[62183]: INFO nova.compute.manager [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 907.659397] env[62183]: DEBUG oslo.service.loopingcall [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.659397] env[62183]: DEBUG nova.compute.manager [-] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.659397] env[62183]: DEBUG nova.network.neutron [-] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 907.684200] env[62183]: DEBUG oslo_concurrency.lockutils [None req-45ae5bf5-7329-4939-981a-5cc1023699d0 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "interface-1b0593f2-b712-4926-bfb8-30c92070ff6a-e7836918-44a9-4db8-8daf-35c2032e5cdb" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.754s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.730375] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387185, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.747572] env[62183]: DEBUG nova.network.neutron [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Successfully created port: f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 907.769766] env[62183]: DEBUG nova.compute.manager [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 907.854260] env[62183]: DEBUG oslo_concurrency.lockutils [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.854586] env[62183]: DEBUG oslo_concurrency.lockutils [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.854781] env[62183]: INFO nova.compute.manager [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Shelving [ 908.064671] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387182, 'name': CloneVM_Task} progress is 95%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.231330] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387185, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.607979} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.231559] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 46aa1062-3fcd-4581-8647-32e1372e9491/46aa1062-3fcd-4581-8647-32e1372e9491.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.231791] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.232063] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71b0e597-df2a-47f0-a1f8-b19fd1825e0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.239057] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 908.239057] env[62183]: value = "task-1387186" [ 908.239057] env[62183]: _type = "Task" [ 908.239057] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.246941] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387186, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.271135] env[62183]: DEBUG nova.compute.manager [req-5d9cebf2-fa28-40b3-90a8-2843f65f81cd req-dd8afc52-b85a-427f-a13a-f2cfc51f537a service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Received event network-vif-deleted-9c452a0f-d3ff-4e17-92b6-a02b07bc1de2 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.271289] env[62183]: INFO nova.compute.manager [req-5d9cebf2-fa28-40b3-90a8-2843f65f81cd req-dd8afc52-b85a-427f-a13a-f2cfc51f537a service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Neutron deleted interface 9c452a0f-d3ff-4e17-92b6-a02b07bc1de2; detaching it from the instance and deleting it from the info cache [ 908.271462] env[62183]: DEBUG nova.network.neutron [req-5d9cebf2-fa28-40b3-90a8-2843f65f81cd req-dd8afc52-b85a-427f-a13a-f2cfc51f537a service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.363894] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.364177] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0b1e9421-cb05-48c4-8721-c36f410a25c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.370949] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 908.370949] env[62183]: value = "task-1387187" [ 908.370949] env[62183]: _type = "Task" [ 908.370949] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.379069] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387187, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.560868] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387182, 'name': CloneVM_Task, 'duration_secs': 2.119525} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.561172] env[62183]: INFO nova.virt.vmwareapi.vmops [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Created linked-clone VM from snapshot [ 908.561951] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a96a02-e1f1-4dac-ba62-a8b6436a3308 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.569138] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Uploading image aa118f5a-60fe-4272-b51b-366d31a6e666 {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 908.595452] env[62183]: DEBUG oslo_vmware.rw_handles [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 908.595452] env[62183]: value = "vm-294474" [ 908.595452] env[62183]: _type = "VirtualMachine" [ 908.595452] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 908.596132] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-899a35fd-c449-4fb6-a3f2-4883653f0220 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.602476] env[62183]: DEBUG oslo_vmware.rw_handles [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease: (returnval){ [ 908.602476] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]520e77c4-8965-5bde-1564-f1763abd45ef" [ 908.602476] env[62183]: _type = "HttpNfcLease" [ 908.602476] env[62183]: } obtained for exporting VM: (result){ [ 908.602476] env[62183]: value = "vm-294474" [ 908.602476] env[62183]: _type = "VirtualMachine" [ 908.602476] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 908.602787] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the lease: (returnval){ [ 908.602787] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]520e77c4-8965-5bde-1564-f1763abd45ef" [ 908.602787] env[62183]: _type = "HttpNfcLease" [ 908.602787] env[62183]: } to be ready. {{(pid=62183) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 908.609331] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 908.609331] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]520e77c4-8965-5bde-1564-f1763abd45ef" [ 908.609331] env[62183]: _type = "HttpNfcLease" [ 908.609331] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 908.639838] env[62183]: DEBUG nova.network.neutron [-] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.749010] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387186, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078733} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.749342] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.750141] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30e3ba8-be8e-4408-936e-1e2eba8eb9ed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.771446] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 46aa1062-3fcd-4581-8647-32e1372e9491/46aa1062-3fcd-4581-8647-32e1372e9491.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.771744] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da0c3ec9-83b5-4e78-b201-16852a51c42a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.787332] env[62183]: DEBUG nova.compute.manager [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 908.789996] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-996ca8a3-e9c4-48bf-a44a-3088df73a175 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.798518] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecb232e-0c78-4a41-be34-17284fc78bf2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.811501] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 908.811501] env[62183]: value = "task-1387189" [ 908.811501] env[62183]: _type = "Task" [ 908.811501] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.818868] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 908.818868] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 908.819131] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.819237] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 908.819385] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.819537] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 908.819746] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 908.819918] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 908.820104] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 908.820302] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 908.820541] env[62183]: DEBUG nova.virt.hardware [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.821384] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59999f95-3795-4b2f-8973-e46446c0bdc5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.835713] env[62183]: DEBUG nova.compute.manager [req-5d9cebf2-fa28-40b3-90a8-2843f65f81cd req-dd8afc52-b85a-427f-a13a-f2cfc51f537a service nova] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Detach interface failed, port_id=9c452a0f-d3ff-4e17-92b6-a02b07bc1de2, reason: Instance 14ea1b33-ffbd-4c48-84ea-443c10d8317d could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 908.836113] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387189, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.836852] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.837122] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.837346] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.837541] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.837711] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.841776] env[62183]: INFO nova.compute.manager [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Terminating instance [ 908.843883] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e136ecc-d56d-4071-a729-292eae78d908 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.848074] env[62183]: DEBUG nova.compute.manager [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.848275] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 908.849221] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92fe18be-eef2-46e2-aa4a-4a8761ebd684 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.863524] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.863774] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ffbd64a-a08f-4880-b341-e4f6862ffcfb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.870342] env[62183]: DEBUG oslo_vmware.api [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 908.870342] env[62183]: value = "task-1387190" [ 908.870342] env[62183]: _type = "Task" [ 908.870342] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.883038] env[62183]: DEBUG oslo_vmware.api [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387190, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.886236] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387187, 'name': PowerOffVM_Task, 'duration_secs': 0.229678} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.886499] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 908.887316] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7b7302-3962-4376-8aa2-ac560b17b4a3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.905169] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86fe1b3d-7762-4ed1-b337-1c052b821ad2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.112056] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 909.112056] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]520e77c4-8965-5bde-1564-f1763abd45ef" [ 909.112056] env[62183]: _type = "HttpNfcLease" [ 909.112056] env[62183]: } is ready. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 909.112056] env[62183]: DEBUG oslo_vmware.rw_handles [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 909.112056] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]520e77c4-8965-5bde-1564-f1763abd45ef" [ 909.112056] env[62183]: _type = "HttpNfcLease" [ 909.112056] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 909.112687] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1518cf8-ef29-4f7a-a3ec-9b08d77f66e6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.119950] env[62183]: DEBUG oslo_vmware.rw_handles [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e231e2-ab32-0d3c-a58d-061bfc24d630/disk-0.vmdk from lease info. {{(pid=62183) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 909.120147] env[62183]: DEBUG oslo_vmware.rw_handles [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e231e2-ab32-0d3c-a58d-061bfc24d630/disk-0.vmdk for reading. {{(pid=62183) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 909.175534] env[62183]: INFO nova.compute.manager [-] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Took 1.52 seconds to deallocate network for instance. [ 909.175948] env[62183]: DEBUG oslo_concurrency.lockutils [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "interface-1b0593f2-b712-4926-bfb8-30c92070ff6a-e7836918-44a9-4db8-8daf-35c2032e5cdb" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.176209] env[62183]: DEBUG oslo_concurrency.lockutils [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "interface-1b0593f2-b712-4926-bfb8-30c92070ff6a-e7836918-44a9-4db8-8daf-35c2032e5cdb" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.209926] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d1e764c0-c145-4e7e-b3ab-26343d9e9241 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.321629] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387189, 'name': ReconfigVM_Task, 'duration_secs': 0.260371} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.321915] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 46aa1062-3fcd-4581-8647-32e1372e9491/46aa1062-3fcd-4581-8647-32e1372e9491.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.322582] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd1e46db-d5fd-4ce7-96da-d7a946f7e74a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.328685] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 909.328685] env[62183]: value = "task-1387191" [ 909.328685] env[62183]: _type = "Task" [ 909.328685] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.336472] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387191, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.382012] env[62183]: DEBUG oslo_vmware.api [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387190, 'name': PowerOffVM_Task, 'duration_secs': 0.235728} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.382408] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.382685] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 909.383106] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2cf82d6a-795d-43b8-98a0-ed44dc1e7829 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.415888] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Creating Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 909.416323] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9fa13fd4-b418-4e0b-9e31-13ec712caf69 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.423356] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 909.423356] env[62183]: value = "task-1387193" [ 909.423356] env[62183]: _type = "Task" [ 909.423356] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.433547] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387193, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.454755] env[62183]: DEBUG nova.network.neutron [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Successfully updated port: f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.464029] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.464141] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.464292] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleting the datastore file [datastore1] 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.465518] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ad02589-5946-41d8-976d-f85ed51e783e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.473973] env[62183]: DEBUG oslo_vmware.api [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 909.473973] env[62183]: value = "task-1387194" [ 909.473973] env[62183]: _type = "Task" [ 909.473973] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.485089] env[62183]: DEBUG oslo_vmware.api [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.679338] env[62183]: DEBUG oslo_concurrency.lockutils [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.680102] env[62183]: DEBUG oslo_concurrency.lockutils [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.680621] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ff7a5d-3df9-4df6-b940-367791729b9b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.686272] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.686639] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.686940] env[62183]: DEBUG nova.objects.instance [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lazy-loading 'resources' on Instance uuid 14ea1b33-ffbd-4c48-84ea-443c10d8317d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.704683] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d118508-f257-43e6-9730-969a427696dd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.733825] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Reconfiguring VM to detach interface {{(pid=62183) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 909.734294] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a03e8c1-2d99-4861-9d97-d8cfdbe477ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.753305] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 909.753305] env[62183]: value = "task-1387195" [ 909.753305] env[62183]: _type = "Task" [ 909.753305] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.761844] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.838588] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387191, 'name': Rename_Task, 'duration_secs': 0.159946} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.839168] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 909.839319] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7acad966-8909-4449-8efa-88676e13b7f2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.845058] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 909.845058] env[62183]: value = "task-1387196" [ 909.845058] env[62183]: _type = "Task" [ 909.845058] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.852747] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387196, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.934107] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387193, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.958041] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.958198] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.958485] env[62183]: DEBUG nova.network.neutron [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 909.984963] env[62183]: DEBUG oslo_vmware.api [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.287416} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.985568] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 909.986279] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 909.986480] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 909.986667] env[62183]: INFO nova.compute.manager [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Took 1.14 seconds to destroy the instance on the hypervisor. [ 909.986930] env[62183]: DEBUG oslo.service.loopingcall [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.987146] env[62183]: DEBUG nova.compute.manager [-] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 909.987266] env[62183]: DEBUG nova.network.neutron [-] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 910.263310] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.304488] env[62183]: DEBUG nova.compute.manager [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Received event network-vif-plugged-f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.304731] env[62183]: DEBUG oslo_concurrency.lockutils [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] Acquiring lock "28d279bd-206e-4ba0-bf84-2980ab41f38d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.305110] env[62183]: DEBUG oslo_concurrency.lockutils [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.305329] env[62183]: DEBUG oslo_concurrency.lockutils [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.305534] env[62183]: DEBUG nova.compute.manager [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] No waiting events found dispatching network-vif-plugged-f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.305708] env[62183]: WARNING nova.compute.manager [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Received unexpected event network-vif-plugged-f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc for instance with vm_state building and task_state spawning. [ 910.305900] env[62183]: DEBUG nova.compute.manager [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Received event network-changed-f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.306113] env[62183]: DEBUG nova.compute.manager [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Refreshing instance network info cache due to event network-changed-f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 910.306344] env[62183]: DEBUG oslo_concurrency.lockutils [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] Acquiring lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.358667] env[62183]: DEBUG oslo_vmware.api [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387196, 'name': PowerOnVM_Task, 'duration_secs': 0.452455} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.359047] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 910.359333] env[62183]: INFO nova.compute.manager [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Took 7.42 seconds to spawn the instance on the hypervisor. [ 910.359675] env[62183]: DEBUG nova.compute.manager [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 910.360492] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cb1798-f934-436f-b60f-c800d933a344 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.401748] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f7d13b-952e-4519-9e8f-735d2b355242 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.411013] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75489598-a18e-4ddf-8f6a-1b3c960c954c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.447393] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d2577d-5b54-4d6c-836c-f2027f0c0ff3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.458296] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387193, 'name': CreateSnapshot_Task, 'duration_secs': 0.784099} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.458756] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Created Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 910.460083] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe3ea5e-13ed-472f-ae45-1da4cb1746bf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.464764] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d783ae15-5f38-44e3-acb9-d6ac14d4ad86 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.482372] env[62183]: DEBUG nova.compute.provider_tree [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.509548] env[62183]: DEBUG nova.network.neutron [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 910.657429] env[62183]: DEBUG nova.network.neutron [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance_info_cache with network_info: [{"id": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "address": "fa:16:3e:39:9c:d2", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2ec12e2-ad", "ovs_interfaceid": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.763097] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.836347] env[62183]: DEBUG nova.network.neutron [-] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.882082] env[62183]: INFO nova.compute.manager [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Took 14.51 seconds to build instance. [ 910.988015] env[62183]: DEBUG nova.scheduler.client.report [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 910.998258] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Creating linked-clone VM from snapshot {{(pid=62183) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 911.000051] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-af379cdc-ac75-4972-8613-a22fd9afc417 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.007614] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 911.007614] env[62183]: value = "task-1387197" [ 911.007614] env[62183]: _type = "Task" [ 911.007614] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.017176] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387197, 'name': CloneVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.160636] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.161091] env[62183]: DEBUG nova.compute.manager [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Instance network_info: |[{"id": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "address": "fa:16:3e:39:9c:d2", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2ec12e2-ad", "ovs_interfaceid": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 911.161512] env[62183]: DEBUG oslo_concurrency.lockutils [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] Acquired lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.161690] env[62183]: DEBUG nova.network.neutron [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Refreshing network info cache for port f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 911.163826] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:9c:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.171826] env[62183]: DEBUG oslo.service.loopingcall [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.172370] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 911.173369] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-79cfe68e-75cf-4109-87ff-d2d69cc4961c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.193843] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.193843] env[62183]: value = "task-1387198" [ 911.193843] env[62183]: _type = "Task" [ 911.193843] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.203251] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387198, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.265183] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.339081] env[62183]: INFO nova.compute.manager [-] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Took 1.35 seconds to deallocate network for instance. [ 911.384647] env[62183]: DEBUG oslo_concurrency.lockutils [None req-617546f4-b986-436b-b862-4f9778591061 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "46aa1062-3fcd-4581-8647-32e1372e9491" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.022s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.503759] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.520806] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387197, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.533123] env[62183]: INFO nova.scheduler.client.report [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted allocations for instance 14ea1b33-ffbd-4c48-84ea-443c10d8317d [ 911.705135] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387198, 'name': CreateVM_Task, 'duration_secs': 0.355566} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.705202] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.706066] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.706066] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.706449] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 911.706648] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c21d86db-4850-4413-b18b-8fc495a63374 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.712238] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 911.712238] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52d3d7fc-bf0a-2d01-ebe5-9fec89b02802" [ 911.712238] env[62183]: _type = "Task" [ 911.712238] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.720087] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d3d7fc-bf0a-2d01-ebe5-9fec89b02802, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.765108] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.847060] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.847988] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.847988] env[62183]: DEBUG nova.objects.instance [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lazy-loading 'resources' on Instance uuid 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.949205] env[62183]: DEBUG nova.network.neutron [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updated VIF entry in instance network info cache for port f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 911.949647] env[62183]: DEBUG nova.network.neutron [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance_info_cache with network_info: [{"id": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "address": "fa:16:3e:39:9c:d2", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2ec12e2-ad", "ovs_interfaceid": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.018921] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387197, 'name': CloneVM_Task} progress is 95%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.040467] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7dfe8f71-522c-46ed-920e-9e636991f5ab tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "14ea1b33-ffbd-4c48-84ea-443c10d8317d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.518s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.224449] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d3d7fc-bf0a-2d01-ebe5-9fec89b02802, 'name': SearchDatastore_Task, 'duration_secs': 0.016685} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.224772] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.225046] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.225303] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.225458] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.226062] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.226374] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c1feb91-8513-4e08-bf17-6cd50a8599b4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.235938] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.236159] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 912.237151] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd8befeb-4086-4ed8-a6ea-62a3e324f2c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.242789] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 912.242789] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]522c0b81-45f1-958b-6174-66e1422e7cfc" [ 912.242789] env[62183]: _type = "Task" [ 912.242789] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.251150] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]522c0b81-45f1-958b-6174-66e1422e7cfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.263765] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.335418] env[62183]: DEBUG nova.compute.manager [req-e6f1b4d7-4808-4fbd-8cef-6ab817b85c31 req-f911b7e3-a4d0-441e-96f2-6007823c3d9b service nova] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Received event network-vif-deleted-43abe649-25d9-41ef-9bd2-d09e3d07f261 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.452277] env[62183]: DEBUG oslo_concurrency.lockutils [req-aa8e4199-530b-44a4-a788-0a3300ff85f1 req-da4f8692-bfa2-4077-8a5d-4e1ff612fd37 service nova] Releasing lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.500460] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16bc847-0201-4a96-a528-1692a3add3f5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.510197] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc61d3d-d575-4c4d-a45f-1ad7b8658bda {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.521519] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387197, 'name': CloneVM_Task, 'duration_secs': 1.145147} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.547396] env[62183]: INFO nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Created linked-clone VM from snapshot [ 912.548647] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370a0c60-5cae-4401-9076-aa98132f0033 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.551682] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c6323d-2c93-4738-9b90-8455742203ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.561349] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Uploading image b9d18b83-05a1-4aa4-87fa-09e4ebd409de {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 912.564644] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e03affc-6257-4f98-a288-4fbee6e3d58e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.578593] env[62183]: DEBUG nova.compute.provider_tree [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.587166] env[62183]: DEBUG oslo_vmware.rw_handles [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 912.587166] env[62183]: value = "vm-294476" [ 912.587166] env[62183]: _type = "VirtualMachine" [ 912.587166] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 912.587461] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4808a36e-d536-4953-881f-6b2cee59fc72 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.595054] env[62183]: DEBUG oslo_vmware.rw_handles [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lease: (returnval){ [ 912.595054] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52172c87-fc5d-5c86-2394-0b2495f608ad" [ 912.595054] env[62183]: _type = "HttpNfcLease" [ 912.595054] env[62183]: } obtained for exporting VM: (result){ [ 912.595054] env[62183]: value = "vm-294476" [ 912.595054] env[62183]: _type = "VirtualMachine" [ 912.595054] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 912.595451] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the lease: (returnval){ [ 912.595451] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52172c87-fc5d-5c86-2394-0b2495f608ad" [ 912.595451] env[62183]: _type = "HttpNfcLease" [ 912.595451] env[62183]: } to be ready. {{(pid=62183) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 912.603141] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 912.603141] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52172c87-fc5d-5c86-2394-0b2495f608ad" [ 912.603141] env[62183]: _type = "HttpNfcLease" [ 912.603141] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 912.752819] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]522c0b81-45f1-958b-6174-66e1422e7cfc, 'name': SearchDatastore_Task, 'duration_secs': 0.010261} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.753660] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-415210c0-132f-4f5c-9f75-857e3c86e7ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.761668] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 912.761668] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f104e6-bc7f-2a74-0590-5607efb00b10" [ 912.761668] env[62183]: _type = "Task" [ 912.761668] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.765320] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.772779] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f104e6-bc7f-2a74-0590-5607efb00b10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.791352] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.791594] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.969046] env[62183]: DEBUG nova.compute.manager [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.970746] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70d8384-8469-465f-af37-31f58fcd1a22 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.083641] env[62183]: DEBUG nova.scheduler.client.report [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 913.106537] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 913.106537] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52172c87-fc5d-5c86-2394-0b2495f608ad" [ 913.106537] env[62183]: _type = "HttpNfcLease" [ 913.106537] env[62183]: } is ready. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 913.106859] env[62183]: DEBUG oslo_vmware.rw_handles [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 913.106859] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52172c87-fc5d-5c86-2394-0b2495f608ad" [ 913.106859] env[62183]: _type = "HttpNfcLease" [ 913.106859] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 913.107648] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de810c9a-55d8-498b-991b-76d4a09020d4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.115619] env[62183]: DEBUG oslo_vmware.rw_handles [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52051f52-52c8-c012-23be-f4c3b782c0ee/disk-0.vmdk from lease info. {{(pid=62183) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 913.115937] env[62183]: DEBUG oslo_vmware.rw_handles [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52051f52-52c8-c012-23be-f4c3b782c0ee/disk-0.vmdk for reading. {{(pid=62183) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 913.223426] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d28df811-35aa-40d1-8ab7-a58180895822 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.267539] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.276577] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f104e6-bc7f-2a74-0590-5607efb00b10, 'name': SearchDatastore_Task, 'duration_secs': 0.012409} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.276896] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.277217] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 28d279bd-206e-4ba0-bf84-2980ab41f38d/28d279bd-206e-4ba0-bf84-2980ab41f38d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 913.277872] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9cb4375-8b9f-41bc-8091-ff5d1ea95787 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.285159] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 913.285159] env[62183]: value = "task-1387200" [ 913.285159] env[62183]: _type = "Task" [ 913.285159] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.295009] env[62183]: DEBUG nova.compute.manager [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 913.298541] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387200, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.484275] env[62183]: INFO nova.compute.manager [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] instance snapshotting [ 913.487814] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37813608-bec8-4c71-910b-a009902fe481 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.509313] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a848a24-b5ba-4e92-8b7b-fa06c9465acc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.589672] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.742s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.618173] env[62183]: INFO nova.scheduler.client.report [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted allocations for instance 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7 [ 913.772526] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.801850] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387200, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.815382] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.815900] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.817592] env[62183]: INFO nova.compute.claims [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.020695] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Creating Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 914.021131] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8c6fc9b8-40ea-4765-afc2-00287c89dc0c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.028758] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 914.028758] env[62183]: value = "task-1387201" [ 914.028758] env[62183]: _type = "Task" [ 914.028758] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.037755] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387201, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.126720] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec32734a-cb9d-4f8b-941c-c262ce189701 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "4de35fd3-7ef8-4a0b-b180-8b36e308e2e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.289s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.270195] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.297948] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387200, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572135} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.297948] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 28d279bd-206e-4ba0-bf84-2980ab41f38d/28d279bd-206e-4ba0-bf84-2980ab41f38d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 914.298727] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 914.299041] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd2a7db5-d406-4c6f-8df5-aa2af1e0919c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.305690] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 914.305690] env[62183]: value = "task-1387202" [ 914.305690] env[62183]: _type = "Task" [ 914.305690] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.313875] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387202, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.539259] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387201, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.768589] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.816133] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387202, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104068} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.817064] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 914.817995] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6b3123-9f73-4ee3-9368-22f615e3cc58 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.845022] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 28d279bd-206e-4ba0-bf84-2980ab41f38d/28d279bd-206e-4ba0-bf84-2980ab41f38d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.846164] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73e19248-09ac-4722-8ce9-7468f724bf10 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.866683] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 914.866683] env[62183]: value = "task-1387203" [ 914.866683] env[62183]: _type = "Task" [ 914.866683] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.874840] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387203, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.995133] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef575228-ec55-4869-8192-0f67ac99819d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.003883] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52cae00-0a58-4746-88c1-92a3528fcc93 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.040325] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766938b0-189b-4a86-9fc3-4d4d7adf7d71 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.051974] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f887b96b-46a7-4484-bee6-17adf24a4482 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.056543] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387201, 'name': CreateSnapshot_Task, 'duration_secs': 0.75917} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.057019] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Created Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 915.058280] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052ab917-56ef-4309-86f3-a33e3d88cdbc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.069752] env[62183]: DEBUG nova.compute.provider_tree [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.276774] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.377614] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387203, 'name': ReconfigVM_Task, 'duration_secs': 0.436233} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.377988] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 28d279bd-206e-4ba0-bf84-2980ab41f38d/28d279bd-206e-4ba0-bf84-2980ab41f38d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.378704] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3f0029d-0cf1-4390-9084-b9f8885dca9b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.385907] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 915.385907] env[62183]: value = "task-1387204" [ 915.385907] env[62183]: _type = "Task" [ 915.385907] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.394700] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387204, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.573863] env[62183]: DEBUG nova.scheduler.client.report [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 915.588356] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Creating linked-clone VM from snapshot {{(pid=62183) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 915.588613] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3fb1271b-5d5d-4569-9912-ca076885dfc1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.598858] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 915.598858] env[62183]: value = "task-1387205" [ 915.598858] env[62183]: _type = "Task" [ 915.598858] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.608353] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387205, 'name': CloneVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.770793] env[62183]: DEBUG oslo_vmware.api [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387195, 'name': ReconfigVM_Task, 'duration_secs': 5.745045} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.771150] env[62183]: DEBUG oslo_concurrency.lockutils [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.771630] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Reconfigured VM to detach interface {{(pid=62183) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 915.896510] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387204, 'name': Rename_Task, 'duration_secs': 0.199427} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.896897] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.897233] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01a75a3d-1709-45f5-82b8-6e983ac92a99 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.903797] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 915.903797] env[62183]: value = "task-1387206" [ 915.903797] env[62183]: _type = "Task" [ 915.903797] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.912888] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.080533] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.264s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.081493] env[62183]: DEBUG nova.compute.manager [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 916.113440] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387205, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.414531] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387206, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.587300] env[62183]: DEBUG nova.compute.utils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.588883] env[62183]: DEBUG nova.compute.manager [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 916.589065] env[62183]: DEBUG nova.network.neutron [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 916.613544] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387205, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.632253] env[62183]: DEBUG nova.policy [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59edae5aed2e4ccbb980bc0973890baf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '594f9c41496146c68921dad4c8d99ec4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.914873] env[62183]: DEBUG oslo_vmware.api [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387206, 'name': PowerOnVM_Task, 'duration_secs': 0.690305} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.915300] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.915437] env[62183]: INFO nova.compute.manager [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Took 8.13 seconds to spawn the instance on the hypervisor. [ 916.915865] env[62183]: DEBUG nova.compute.manager [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.916478] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e0fee9-8ea4-4415-9413-380e303c7bf6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.925861] env[62183]: DEBUG nova.network.neutron [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Successfully created port: b1af05d2-9221-4ac1-a4b4-82f7013cde5c {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 917.093040] env[62183]: DEBUG nova.compute.manager [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 917.114749] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387205, 'name': CloneVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.138029] env[62183]: DEBUG oslo_concurrency.lockutils [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.138504] env[62183]: DEBUG oslo_concurrency.lockutils [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquired lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.138979] env[62183]: DEBUG nova.network.neutron [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.242781] env[62183]: DEBUG oslo_vmware.rw_handles [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e231e2-ab32-0d3c-a58d-061bfc24d630/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 917.243913] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066b6f31-0f85-47ec-891e-105ab99f78ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.251255] env[62183]: DEBUG oslo_vmware.rw_handles [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e231e2-ab32-0d3c-a58d-061bfc24d630/disk-0.vmdk is in state: ready. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 917.251454] env[62183]: ERROR oslo_vmware.rw_handles [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e231e2-ab32-0d3c-a58d-061bfc24d630/disk-0.vmdk due to incomplete transfer. [ 917.251708] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9a7721e1-031e-461d-aeff-52688ece16f4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.260314] env[62183]: DEBUG oslo_vmware.rw_handles [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e231e2-ab32-0d3c-a58d-061bfc24d630/disk-0.vmdk. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 917.260314] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Uploaded image aa118f5a-60fe-4272-b51b-366d31a6e666 to the Glance image server {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 917.262786] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Destroying the VM {{(pid=62183) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 917.263151] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c34bf64f-94b5-4856-b652-97bdef8bfd5d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.269401] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 917.269401] env[62183]: value = "task-1387207" [ 917.269401] env[62183]: _type = "Task" [ 917.269401] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.278950] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387207, 'name': Destroy_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.436141] env[62183]: INFO nova.compute.manager [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Took 14.14 seconds to build instance. [ 917.596684] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.597055] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.597287] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "1b0593f2-b712-4926-bfb8-30c92070ff6a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.597484] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.597664] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.604778] env[62183]: INFO nova.compute.manager [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Terminating instance [ 917.607295] env[62183]: DEBUG nova.compute.manager [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 917.607413] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.608436] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98211f9-76c0-452b-ba6f-fce1393091ac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.623025] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.627344] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d58620e-06aa-4f97-b936-735663d12059 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.629213] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387205, 'name': CloneVM_Task, 'duration_secs': 1.519183} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.629481] env[62183]: INFO nova.virt.vmwareapi.vmops [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Created linked-clone VM from snapshot [ 917.630682] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818f2316-2ebf-4f0b-8f55-02787079b605 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.634559] env[62183]: DEBUG oslo_vmware.api [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 917.634559] env[62183]: value = "task-1387208" [ 917.634559] env[62183]: _type = "Task" [ 917.634559] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.643720] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Uploading image 634230f0-6c36-4cb3-8e4e-ed6f18ee42e6 {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 917.653454] env[62183]: DEBUG oslo_vmware.api [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387208, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.657232] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Destroying the VM {{(pid=62183) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 917.658066] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c3371694-70a5-47be-9762-de591337245f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.665079] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 917.665079] env[62183]: value = "task-1387209" [ 917.665079] env[62183]: _type = "Task" [ 917.665079] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.675496] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387209, 'name': Destroy_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.779990] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387207, 'name': Destroy_Task} progress is 33%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.912099] env[62183]: INFO nova.network.neutron [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Port e7836918-44a9-4db8-8daf-35c2032e5cdb from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 917.912601] env[62183]: DEBUG nova.network.neutron [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updating instance_info_cache with network_info: [{"id": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "address": "fa:16:3e:46:24:80", "network": {"id": "ece15cf2-6921-4b76-9d7e-f313bcaa2f48", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-100692171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ebcc716f958942b588a6bfde78d2c00d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap999bf20b-d0", "ovs_interfaceid": "999bf20b-d0e4-4229-a150-ec7c31e38cc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.938683] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5bd14e8d-2556-4a20-a443-a1d492b49d25 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.647s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.096493] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.096740] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.105882] env[62183]: DEBUG nova.compute.manager [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 918.132396] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.132665] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.132826] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.133030] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.133195] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.133347] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.133562] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.133728] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.133897] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.134076] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.134270] env[62183]: DEBUG nova.virt.hardware [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.135174] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b1f249-b342-4afd-8814-dbe440ac0288 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.146622] env[62183]: DEBUG oslo_vmware.api [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387208, 'name': PowerOffVM_Task, 'duration_secs': 0.310594} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.148849] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.149055] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.149367] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-58ef3979-dbcc-4503-88e2-d05f1f4a4d0a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.151855] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f03298-7085-4bc5-b94b-b35a6d40e694 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.174366] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387209, 'name': Destroy_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.219928] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.220193] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.220384] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Deleting the datastore file [datastore1] 1b0593f2-b712-4926-bfb8-30c92070ff6a {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.220673] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6c9c36d-7b76-4761-8023-27b30a5a2cb0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.227557] env[62183]: DEBUG oslo_vmware.api [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 918.227557] env[62183]: value = "task-1387211" [ 918.227557] env[62183]: _type = "Task" [ 918.227557] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.236124] env[62183]: DEBUG oslo_vmware.api [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387211, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.282037] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387207, 'name': Destroy_Task, 'duration_secs': 0.673231} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.282319] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Destroyed the VM [ 918.282920] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deleting Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 918.283449] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0ff75700-31a9-4b35-8ac2-62c23955c7f9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.290495] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 918.290495] env[62183]: value = "task-1387212" [ 918.290495] env[62183]: _type = "Task" [ 918.290495] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.300774] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387212, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.400219] env[62183]: DEBUG nova.compute.manager [req-d386debd-d408-4559-a6c2-1c3d10f93fbc req-3825b877-5c47-449a-a488-4f5047b867ce service nova] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Received event network-vif-plugged-b1af05d2-9221-4ac1-a4b4-82f7013cde5c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.400435] env[62183]: DEBUG oslo_concurrency.lockutils [req-d386debd-d408-4559-a6c2-1c3d10f93fbc req-3825b877-5c47-449a-a488-4f5047b867ce service nova] Acquiring lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.400681] env[62183]: DEBUG oslo_concurrency.lockutils [req-d386debd-d408-4559-a6c2-1c3d10f93fbc req-3825b877-5c47-449a-a488-4f5047b867ce service nova] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.400895] env[62183]: DEBUG oslo_concurrency.lockutils [req-d386debd-d408-4559-a6c2-1c3d10f93fbc req-3825b877-5c47-449a-a488-4f5047b867ce service nova] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.401116] env[62183]: DEBUG nova.compute.manager [req-d386debd-d408-4559-a6c2-1c3d10f93fbc req-3825b877-5c47-449a-a488-4f5047b867ce service nova] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] No waiting events found dispatching network-vif-plugged-b1af05d2-9221-4ac1-a4b4-82f7013cde5c {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.401309] env[62183]: WARNING nova.compute.manager [req-d386debd-d408-4559-a6c2-1c3d10f93fbc req-3825b877-5c47-449a-a488-4f5047b867ce service nova] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Received unexpected event network-vif-plugged-b1af05d2-9221-4ac1-a4b4-82f7013cde5c for instance with vm_state building and task_state spawning. [ 918.415429] env[62183]: DEBUG oslo_concurrency.lockutils [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Releasing lock "refresh_cache-1b0593f2-b712-4926-bfb8-30c92070ff6a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.570612] env[62183]: DEBUG nova.network.neutron [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Successfully updated port: b1af05d2-9221-4ac1-a4b4-82f7013cde5c {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.599620] env[62183]: DEBUG nova.compute.manager [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 918.677389] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387209, 'name': Destroy_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.737728] env[62183]: DEBUG oslo_vmware.api [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387211, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265465} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.738025] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.738225] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.738488] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.738711] env[62183]: INFO nova.compute.manager [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 918.738969] env[62183]: DEBUG oslo.service.loopingcall [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.739183] env[62183]: DEBUG nova.compute.manager [-] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 918.739280] env[62183]: DEBUG nova.network.neutron [-] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 918.800013] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387212, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.919377] env[62183]: DEBUG oslo_concurrency.lockutils [None req-19c517cf-a499-4041-9409-ff6c5d038ac2 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "interface-1b0593f2-b712-4926-bfb8-30c92070ff6a-e7836918-44a9-4db8-8daf-35c2032e5cdb" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.743s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.072564] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.072899] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.073040] env[62183]: DEBUG nova.network.neutron [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 919.119946] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.120598] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.121860] env[62183]: INFO nova.compute.claims [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.185295] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387209, 'name': Destroy_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.301349] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387212, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.616769] env[62183]: DEBUG nova.network.neutron [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 919.686194] env[62183]: DEBUG nova.compute.manager [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Stashing vm_state: active {{(pid=62183) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 919.698845] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387209, 'name': Destroy_Task, 'duration_secs': 1.525328} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.699122] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Destroyed the VM [ 919.699469] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Deleting Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 919.699819] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3a0b845e-c4ec-4062-b7fa-e638f555f57d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.709819] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 919.709819] env[62183]: value = "task-1387213" [ 919.709819] env[62183]: _type = "Task" [ 919.709819] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.722159] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387213, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.798281] env[62183]: DEBUG nova.network.neutron [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance_info_cache with network_info: [{"id": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "address": "fa:16:3e:a4:03:ac", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1af05d2-92", "ovs_interfaceid": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.806079] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387212, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.068479] env[62183]: DEBUG nova.network.neutron [-] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.207773] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.225781] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387213, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.303559] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.304165] env[62183]: DEBUG nova.compute.manager [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Instance network_info: |[{"id": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "address": "fa:16:3e:a4:03:ac", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1af05d2-92", "ovs_interfaceid": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.305012] env[62183]: DEBUG oslo_vmware.api [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387212, 'name': RemoveSnapshot_Task, 'duration_secs': 1.527666} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.311036] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:03:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '496ac502-bfc4-4324-8332-cac473eb7cc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1af05d2-9221-4ac1-a4b4-82f7013cde5c', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.319619] env[62183]: DEBUG oslo.service.loopingcall [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.319990] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deleted Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 920.320878] env[62183]: INFO nova.compute.manager [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Took 15.90 seconds to snapshot the instance on the hypervisor. [ 920.325944] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 920.327584] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f751e95f-73aa-4533-bb19-aad27e23b257 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.350322] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.350322] env[62183]: value = "task-1387214" [ 920.350322] env[62183]: _type = "Task" [ 920.350322] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.359230] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387214, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.364576] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8654f235-6be8-4658-afde-e5f26570d7e8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.372539] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b609fc4a-212b-406b-acef-cf7b329d218a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.411835] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6dfb60-bf99-43ec-ad02-1c39076254a0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.420500] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239e45c6-cb5d-4670-9696-89ee645aa89f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.436687] env[62183]: DEBUG nova.compute.provider_tree [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.440135] env[62183]: DEBUG nova.compute.manager [req-5b32ba7f-6213-474e-9c9b-1ead3e2029dc req-b81faf0b-369e-4711-ba54-3946184fb8e3 service nova] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Received event network-changed-b1af05d2-9221-4ac1-a4b4-82f7013cde5c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.440408] env[62183]: DEBUG nova.compute.manager [req-5b32ba7f-6213-474e-9c9b-1ead3e2029dc req-b81faf0b-369e-4711-ba54-3946184fb8e3 service nova] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Refreshing instance network info cache due to event network-changed-b1af05d2-9221-4ac1-a4b4-82f7013cde5c. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 920.440707] env[62183]: DEBUG oslo_concurrency.lockutils [req-5b32ba7f-6213-474e-9c9b-1ead3e2029dc req-b81faf0b-369e-4711-ba54-3946184fb8e3 service nova] Acquiring lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.440915] env[62183]: DEBUG oslo_concurrency.lockutils [req-5b32ba7f-6213-474e-9c9b-1ead3e2029dc req-b81faf0b-369e-4711-ba54-3946184fb8e3 service nova] Acquired lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.441163] env[62183]: DEBUG nova.network.neutron [req-5b32ba7f-6213-474e-9c9b-1ead3e2029dc req-b81faf0b-369e-4711-ba54-3946184fb8e3 service nova] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Refreshing network info cache for port b1af05d2-9221-4ac1-a4b4-82f7013cde5c {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.571557] env[62183]: INFO nova.compute.manager [-] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Took 1.83 seconds to deallocate network for instance. [ 920.720562] env[62183]: DEBUG oslo_vmware.api [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387213, 'name': RemoveSnapshot_Task, 'duration_secs': 0.842779} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.720874] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Deleted Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 920.860204] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387214, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.945344] env[62183]: DEBUG nova.scheduler.client.report [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 921.078893] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.085404] env[62183]: DEBUG nova.compute.manager [None req-5fff16df-42db-48dc-a04a-18ca683c3c6d tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Found 2 images (rotation: 2) {{(pid=62183) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 921.168687] env[62183]: DEBUG nova.network.neutron [req-5b32ba7f-6213-474e-9c9b-1ead3e2029dc req-b81faf0b-369e-4711-ba54-3946184fb8e3 service nova] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updated VIF entry in instance network info cache for port b1af05d2-9221-4ac1-a4b4-82f7013cde5c. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.169085] env[62183]: DEBUG nova.network.neutron [req-5b32ba7f-6213-474e-9c9b-1ead3e2029dc req-b81faf0b-369e-4711-ba54-3946184fb8e3 service nova] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance_info_cache with network_info: [{"id": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "address": "fa:16:3e:a4:03:ac", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1af05d2-92", "ovs_interfaceid": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.178315] env[62183]: DEBUG oslo_vmware.rw_handles [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52051f52-52c8-c012-23be-f4c3b782c0ee/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 921.179288] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c608e0-8cd1-4301-96b9-f26ddeeaa247 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.185962] env[62183]: DEBUG oslo_vmware.rw_handles [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52051f52-52c8-c012-23be-f4c3b782c0ee/disk-0.vmdk is in state: ready. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 921.186151] env[62183]: ERROR oslo_vmware.rw_handles [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52051f52-52c8-c012-23be-f4c3b782c0ee/disk-0.vmdk due to incomplete transfer. [ 921.186668] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-73739d65-f6e1-4d36-a206-06c8f8851ddd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.194188] env[62183]: DEBUG oslo_vmware.rw_handles [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52051f52-52c8-c012-23be-f4c3b782c0ee/disk-0.vmdk. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 921.194389] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Uploaded image b9d18b83-05a1-4aa4-87fa-09e4ebd409de to the Glance image server {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 921.196130] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Destroying the VM {{(pid=62183) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 921.196394] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fa36b032-cda8-4797-8e3f-8958357775db {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.201946] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 921.201946] env[62183]: value = "task-1387215" [ 921.201946] env[62183]: _type = "Task" [ 921.201946] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.211016] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387215, 'name': Destroy_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.226544] env[62183]: WARNING nova.compute.manager [None req-e23b4c21-4d48-4f98-892d-ac0db9952191 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Image not found during snapshot: nova.exception.ImageNotFound: Image 634230f0-6c36-4cb3-8e4e-ed6f18ee42e6 could not be found. [ 921.362336] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387214, 'name': CreateVM_Task, 'duration_secs': 0.590883} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.362637] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 921.363413] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.363616] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.363965] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.364260] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b573f343-2968-4954-9807-6eab72c7f334 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.369232] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 921.369232] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]520fcb84-5aac-948a-431d-a20ef1b69707" [ 921.369232] env[62183]: _type = "Task" [ 921.369232] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.377131] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]520fcb84-5aac-948a-431d-a20ef1b69707, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.451390] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.451959] env[62183]: DEBUG nova.compute.manager [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 921.455027] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.248s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.527615] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.527897] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.594332] env[62183]: DEBUG nova.compute.manager [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.595387] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e1216e-08d0-4759-aedc-3b04184650cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.674029] env[62183]: DEBUG oslo_concurrency.lockutils [req-5b32ba7f-6213-474e-9c9b-1ead3e2029dc req-b81faf0b-369e-4711-ba54-3946184fb8e3 service nova] Releasing lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.674029] env[62183]: DEBUG nova.compute.manager [req-5b32ba7f-6213-474e-9c9b-1ead3e2029dc req-b81faf0b-369e-4711-ba54-3946184fb8e3 service nova] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Received event network-vif-deleted-999bf20b-d0e4-4229-a150-ec7c31e38cc6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.711866] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387215, 'name': Destroy_Task, 'duration_secs': 0.303096} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.712150] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Destroyed the VM [ 921.712385] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Deleting Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 921.712699] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-19a71331-fb4a-46d0-a1a4-6cd6195e88fa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.719131] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 921.719131] env[62183]: value = "task-1387216" [ 921.719131] env[62183]: _type = "Task" [ 921.719131] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.726932] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387216, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.880073] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]520fcb84-5aac-948a-431d-a20ef1b69707, 'name': SearchDatastore_Task, 'duration_secs': 0.011048} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.880425] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.880675] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.880939] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.881122] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.881330] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.881604] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ecef67d0-b49b-4723-8ea3-07f496dd39fc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.888979] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.889201] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.889944] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2857a54-eec1-4c0f-b391-c5e2cf22a1ba {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.894898] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 921.894898] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52d6c1a7-13e7-87d9-57d8-7667a7ce41a3" [ 921.894898] env[62183]: _type = "Task" [ 921.894898] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.902453] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d6c1a7-13e7-87d9-57d8-7667a7ce41a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.958155] env[62183]: DEBUG nova.compute.utils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 921.961630] env[62183]: INFO nova.compute.claims [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.965924] env[62183]: DEBUG nova.compute.manager [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 921.966145] env[62183]: DEBUG nova.network.neutron [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 922.007018] env[62183]: DEBUG nova.policy [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88efcb2057e84413806f9ed809c260e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eda7e297c551407295b3c67fa0f1dbc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.030653] env[62183]: DEBUG nova.compute.utils [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.077719] env[62183]: DEBUG oslo_concurrency.lockutils [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "46aa1062-3fcd-4581-8647-32e1372e9491" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.077985] env[62183]: DEBUG oslo_concurrency.lockutils [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "46aa1062-3fcd-4581-8647-32e1372e9491" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.078239] env[62183]: DEBUG oslo_concurrency.lockutils [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "46aa1062-3fcd-4581-8647-32e1372e9491-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.078447] env[62183]: DEBUG oslo_concurrency.lockutils [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "46aa1062-3fcd-4581-8647-32e1372e9491-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.078620] env[62183]: DEBUG oslo_concurrency.lockutils [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "46aa1062-3fcd-4581-8647-32e1372e9491-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.080700] env[62183]: INFO nova.compute.manager [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Terminating instance [ 922.082467] env[62183]: DEBUG nova.compute.manager [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.082670] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.083528] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f4b0322-7c86-4cd3-b575-7e3f5a6cc4ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.090842] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.091101] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77c1b18f-5e11-4396-8222-c3c7c2d7af86 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.097790] env[62183]: DEBUG oslo_vmware.api [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 922.097790] env[62183]: value = "task-1387217" [ 922.097790] env[62183]: _type = "Task" [ 922.097790] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.105173] env[62183]: DEBUG oslo_vmware.api [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387217, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.110455] env[62183]: INFO nova.compute.manager [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] instance snapshotting [ 922.110455] env[62183]: DEBUG nova.objects.instance [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'flavor' on Instance uuid a948464b-63aa-4bc8-9885-228049e96d37 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.229303] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387216, 'name': RemoveSnapshot_Task, 'duration_secs': 0.325448} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.229630] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Deleted Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 922.230452] env[62183]: DEBUG nova.compute.manager [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.230680] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9f5aff-1eeb-4ce5-954f-e4bd3e58e4c9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.264150] env[62183]: DEBUG nova.network.neutron [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Successfully created port: ef8e8a41-f9ce-44f2-bcea-feef82484547 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.405407] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d6c1a7-13e7-87d9-57d8-7667a7ce41a3, 'name': SearchDatastore_Task, 'duration_secs': 0.007631} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.406163] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cdff4c1-bc78-43bd-a9ff-35ea4cee4c94 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.411068] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 922.411068] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52d7a250-f48c-1c0c-4e43-0b71297ba1bd" [ 922.411068] env[62183]: _type = "Task" [ 922.411068] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.418294] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d7a250-f48c-1c0c-4e43-0b71297ba1bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.467344] env[62183]: DEBUG nova.compute.manager [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 922.471286] env[62183]: INFO nova.compute.resource_tracker [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating resource usage from migration b2bb67a2-95bd-43aa-8d40-b021bec5cd50 [ 922.534235] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.607655] env[62183]: DEBUG oslo_vmware.api [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387217, 'name': PowerOffVM_Task, 'duration_secs': 0.22546} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.607937] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.608123] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 922.608410] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5abb6350-0780-4707-9051-16388d9da283 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.616395] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209a8970-051c-4de5-bf5d-2c8f3f97f17a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.635050] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc70788-e1e0-42d5-b9b2-802cd757ee9e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.637515] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2beb0a-f679-4ec3-a8db-d5200394da06 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.649040] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421259d1-0b66-488b-9f68-604f10b433e4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.681695] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e301b496-39a0-496c-b818-c3a188d8d777 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.684271] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 922.684512] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 922.684699] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleting the datastore file [datastore1] 46aa1062-3fcd-4581-8647-32e1372e9491 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.685386] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7dbc6bd8-34f9-480e-91d5-c91bfcd99af3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.692458] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edfc6137-ec46-43d9-8fd9-e4e7b3089f5e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.696104] env[62183]: DEBUG oslo_vmware.api [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for the task: (returnval){ [ 922.696104] env[62183]: value = "task-1387219" [ 922.696104] env[62183]: _type = "Task" [ 922.696104] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.707367] env[62183]: DEBUG nova.compute.provider_tree [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.714870] env[62183]: DEBUG oslo_vmware.api [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387219, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.743456] env[62183]: INFO nova.compute.manager [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Shelve offloading [ 922.745067] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.745379] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-699e9ea0-bcb1-4c74-8395-a53ecc7631e1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.751519] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 922.751519] env[62183]: value = "task-1387220" [ 922.751519] env[62183]: _type = "Task" [ 922.751519] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.759254] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387220, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.921463] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d7a250-f48c-1c0c-4e43-0b71297ba1bd, 'name': SearchDatastore_Task, 'duration_secs': 0.009007} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.921754] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.922083] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] b89455d9-6d00-41ac-95ec-fb036b8c26c9/b89455d9-6d00-41ac-95ec-fb036b8c26c9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 922.922463] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d63f73c3-576b-42b9-9670-312da1a5ac89 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.928492] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 922.928492] env[62183]: value = "task-1387221" [ 922.928492] env[62183]: _type = "Task" [ 922.928492] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.936859] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387221, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.149486] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Creating Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 923.149835] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-94b8572e-9609-48db-affd-1221804325f1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.157911] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 923.157911] env[62183]: value = "task-1387222" [ 923.157911] env[62183]: _type = "Task" [ 923.157911] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.166613] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387222, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.207019] env[62183]: DEBUG oslo_vmware.api [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Task: {'id': task-1387219, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129252} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.207331] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.207526] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 923.207709] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 923.207893] env[62183]: INFO nova.compute.manager [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Took 1.13 seconds to destroy the instance on the hypervisor. [ 923.208176] env[62183]: DEBUG oslo.service.loopingcall [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.208391] env[62183]: DEBUG nova.compute.manager [-] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.208487] env[62183]: DEBUG nova.network.neutron [-] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 923.212210] env[62183]: DEBUG nova.scheduler.client.report [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 923.263840] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] VM already powered off {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 923.264290] env[62183]: DEBUG nova.compute.manager [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.265050] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bfd868-4aa3-418d-b20f-437588909bc0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.272958] env[62183]: DEBUG oslo_concurrency.lockutils [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.273305] env[62183]: DEBUG oslo_concurrency.lockutils [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.273452] env[62183]: DEBUG nova.network.neutron [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.440265] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387221, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502827} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.440560] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] b89455d9-6d00-41ac-95ec-fb036b8c26c9/b89455d9-6d00-41ac-95ec-fb036b8c26c9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 923.440832] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 923.441128] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d7290f1-59e9-4d8e-a0b2-f6565735d80f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.447731] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 923.447731] env[62183]: value = "task-1387223" [ 923.447731] env[62183]: _type = "Task" [ 923.447731] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.455446] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387223, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.479845] env[62183]: DEBUG nova.compute.manager [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 923.503222] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.503519] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.503740] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.504096] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.504344] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.504537] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.504815] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.505067] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.505302] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.505492] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.505676] env[62183]: DEBUG nova.virt.hardware [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.506589] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb634818-20de-4b63-9351-b50e5de4b5e7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.514121] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a68a501-5d6c-4205-9a6e-6cc6d727bce0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.595436] env[62183]: DEBUG nova.compute.manager [req-56944953-9eb0-4f57-bc95-952c48132a67 req-ad5edfac-c950-42de-8c56-82dcf025a266 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Received event network-vif-deleted-34118d05-2bab-4122-82fa-0a5ae337cbbe {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.595649] env[62183]: INFO nova.compute.manager [req-56944953-9eb0-4f57-bc95-952c48132a67 req-ad5edfac-c950-42de-8c56-82dcf025a266 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Neutron deleted interface 34118d05-2bab-4122-82fa-0a5ae337cbbe; detaching it from the instance and deleting it from the info cache [ 923.595829] env[62183]: DEBUG nova.network.neutron [req-56944953-9eb0-4f57-bc95-952c48132a67 req-ad5edfac-c950-42de-8c56-82dcf025a266 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.608910] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.609183] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.609528] env[62183]: INFO nova.compute.manager [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Attaching volume f481c3e9-7f36-4566-9973-b98f2c584ee9 to /dev/sdb [ 923.653208] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0a09be-217a-4855-9c66-01bad868b33e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.663164] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723880ef-29dd-4120-9d60-e94b89a117f5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.670405] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387222, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.676552] env[62183]: DEBUG nova.virt.block_device [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Updating existing volume attachment record: 7dd86ba9-03ef-4ac3-8d09-7a575c79173c {{(pid=62183) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 923.717062] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.262s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.717319] env[62183]: INFO nova.compute.manager [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Migrating [ 923.723991] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.645s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.724253] env[62183]: DEBUG nova.objects.instance [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lazy-loading 'resources' on Instance uuid 1b0593f2-b712-4926-bfb8-30c92070ff6a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.801809] env[62183]: DEBUG nova.network.neutron [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Successfully updated port: ef8e8a41-f9ce-44f2-bcea-feef82484547 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.956418] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387223, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063791} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.956752] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.957557] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d90ef32-b98a-4764-b53b-b2db14d75fe9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.979341] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] b89455d9-6d00-41ac-95ec-fb036b8c26c9/b89455d9-6d00-41ac-95ec-fb036b8c26c9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.979676] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b0b53eb-de0c-4b1f-b364-d4ce03763b20 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.000140] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 924.000140] env[62183]: value = "task-1387227" [ 924.000140] env[62183]: _type = "Task" [ 924.000140] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.008928] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387227, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.029784] env[62183]: DEBUG nova.network.neutron [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updating instance_info_cache with network_info: [{"id": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "address": "fa:16:3e:3f:c5:6f", "network": {"id": "7d52241a-8c3c-46bc-b64d-1cbc74ec2434", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-104877467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e74d26cb8f5a47a48dc025b8058c5768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fafdd3-76", "ovs_interfaceid": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.067785] env[62183]: DEBUG nova.network.neutron [-] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.098860] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f58f4f11-281c-4494-a4c2-6d5bea687c41 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.108117] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4e749f-6b83-4ad5-bf91-4ac359642550 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.136309] env[62183]: DEBUG nova.compute.manager [req-56944953-9eb0-4f57-bc95-952c48132a67 req-ad5edfac-c950-42de-8c56-82dcf025a266 service nova] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Detach interface failed, port_id=34118d05-2bab-4122-82fa-0a5ae337cbbe, reason: Instance 46aa1062-3fcd-4581-8647-32e1372e9491 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 924.168919] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387222, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.232322] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.232524] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.232668] env[62183]: DEBUG nova.network.neutron [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.304961] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "refresh_cache-7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.305321] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "refresh_cache-7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.305321] env[62183]: DEBUG nova.network.neutron [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.390845] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a95f549-6b29-465d-adc5-d1213e753cde {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.398821] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f980ef6-9268-45e0-a55f-353d3ca2374e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.429752] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0afab6f-fea7-4075-b4cb-4ae1d3052b63 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.437630] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e504af-ea72-4666-86ff-9150dcf5bcd3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.451567] env[62183]: DEBUG nova.compute.provider_tree [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.510123] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387227, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.532446] env[62183]: DEBUG oslo_concurrency.lockutils [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.571044] env[62183]: INFO nova.compute.manager [-] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Took 1.36 seconds to deallocate network for instance. [ 924.672119] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387222, 'name': CreateSnapshot_Task, 'duration_secs': 1.175192} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.672412] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Created Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 924.673282] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebca964c-2390-4d9a-9dcc-43653479e18b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.855009] env[62183]: DEBUG nova.network.neutron [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 924.954938] env[62183]: DEBUG nova.scheduler.client.report [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 924.984336] env[62183]: DEBUG nova.network.neutron [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Updating instance_info_cache with network_info: [{"id": "ef8e8a41-f9ce-44f2-bcea-feef82484547", "address": "fa:16:3e:db:a5:ae", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef8e8a41-f9", "ovs_interfaceid": "ef8e8a41-f9ce-44f2-bcea-feef82484547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.986307] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 924.987408] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51ebf49-7710-4e60-b935-ae281156fa57 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.995483] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.995675] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3007f0ea-240b-4c53-b0c5-7d029298946c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.010356] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387227, 'name': ReconfigVM_Task, 'duration_secs': 0.708393} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.010356] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Reconfigured VM instance instance-00000055 to attach disk [datastore1] b89455d9-6d00-41ac-95ec-fb036b8c26c9/b89455d9-6d00-41ac-95ec-fb036b8c26c9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 925.010914] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fcddd150-9416-4e06-afb0-dfeb3f6c00ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.017551] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 925.017551] env[62183]: value = "task-1387229" [ 925.017551] env[62183]: _type = "Task" [ 925.017551] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.027299] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387229, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.046942] env[62183]: DEBUG nova.network.neutron [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance_info_cache with network_info: [{"id": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "address": "fa:16:3e:39:9c:d2", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2ec12e2-ad", "ovs_interfaceid": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.077273] env[62183]: DEBUG oslo_concurrency.lockutils [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.092907] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.093247] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.093449] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Deleting the datastore file [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.093787] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-853556cb-ef89-4270-a13c-3390fa431504 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.100806] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 925.100806] env[62183]: value = "task-1387230" [ 925.100806] env[62183]: _type = "Task" [ 925.100806] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.109937] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387230, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.190763] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Creating linked-clone VM from snapshot {{(pid=62183) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 925.191648] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3d867c98-f86d-4ee3-a6a4-c079c3116465 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.199611] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 925.199611] env[62183]: value = "task-1387231" [ 925.199611] env[62183]: _type = "Task" [ 925.199611] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.207990] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387231, 'name': CloneVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.460034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.736s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.462397] env[62183]: DEBUG oslo_concurrency.lockutils [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.385s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.462555] env[62183]: DEBUG nova.objects.instance [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lazy-loading 'resources' on Instance uuid 46aa1062-3fcd-4581-8647-32e1372e9491 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.481755] env[62183]: INFO nova.scheduler.client.report [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Deleted allocations for instance 1b0593f2-b712-4926-bfb8-30c92070ff6a [ 925.491451] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "refresh_cache-7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.491774] env[62183]: DEBUG nova.compute.manager [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Instance network_info: |[{"id": "ef8e8a41-f9ce-44f2-bcea-feef82484547", "address": "fa:16:3e:db:a5:ae", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef8e8a41-f9", "ovs_interfaceid": "ef8e8a41-f9ce-44f2-bcea-feef82484547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 925.492249] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:a5:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef8e8a41-f9ce-44f2-bcea-feef82484547', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 925.500345] env[62183]: DEBUG oslo.service.loopingcall [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.500682] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 925.501302] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ddf62aa8-b7b7-430e-9c3e-abc0fe269e41 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.522444] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 925.522444] env[62183]: value = "task-1387232" [ 925.522444] env[62183]: _type = "Task" [ 925.522444] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.529226] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387229, 'name': Rename_Task, 'duration_secs': 0.138164} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.530649] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.530930] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1fffa8df-e561-49a9-b39b-6593c18a0c15 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.535544] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387232, 'name': CreateVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.539832] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 925.539832] env[62183]: value = "task-1387233" [ 925.539832] env[62183]: _type = "Task" [ 925.539832] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.547357] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.550023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.613255] env[62183]: DEBUG oslo_vmware.api [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387230, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128157} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.613542] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.613746] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.614374] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.625438] env[62183]: DEBUG nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Received event network-vif-plugged-ef8e8a41-f9ce-44f2-bcea-feef82484547 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.625723] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Acquiring lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.625945] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.626335] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.626554] env[62183]: DEBUG nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] No waiting events found dispatching network-vif-plugged-ef8e8a41-f9ce-44f2-bcea-feef82484547 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.626766] env[62183]: WARNING nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Received unexpected event network-vif-plugged-ef8e8a41-f9ce-44f2-bcea-feef82484547 for instance with vm_state building and task_state spawning. [ 925.626972] env[62183]: DEBUG nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Received event network-changed-ef8e8a41-f9ce-44f2-bcea-feef82484547 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.627149] env[62183]: DEBUG nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Refreshing instance network info cache due to event network-changed-ef8e8a41-f9ce-44f2-bcea-feef82484547. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.627396] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Acquiring lock "refresh_cache-7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.627562] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Acquired lock "refresh_cache-7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.627727] env[62183]: DEBUG nova.network.neutron [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Refreshing network info cache for port ef8e8a41-f9ce-44f2-bcea-feef82484547 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 925.632743] env[62183]: INFO nova.scheduler.client.report [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Deleted allocations for instance fb348784-62a0-4d1f-ac7f-f176f3da0dd9 [ 925.709690] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387231, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.990162] env[62183]: DEBUG oslo_concurrency.lockutils [None req-bc02ae9a-08d2-4064-8eb8-11c8bb2239fd tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "1b0593f2-b712-4926-bfb8-30c92070ff6a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.393s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.034711] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387232, 'name': CreateVM_Task, 'duration_secs': 0.359356} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.034911] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 926.035671] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.035865] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.036226] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 926.036550] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c4a179d-b5c8-4045-99da-fff476c5c526 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.041708] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 926.041708] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]526359a1-e1f2-39fa-3825-80c86d563d7f" [ 926.041708] env[62183]: _type = "Task" [ 926.041708] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.058922] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387233, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.065246] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526359a1-e1f2-39fa-3825-80c86d563d7f, 'name': SearchDatastore_Task, 'duration_secs': 0.009986} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.065744] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.065989] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.066295] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.066410] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.066587] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.066853] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-426a80db-837e-4ee5-af29-f9ce96846c05 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.075429] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.075671] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 926.079082] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aad31876-4bef-4aad-b74f-9930d84be6ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.086019] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 926.086019] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5248eb77-b982-4403-ab36-9499b422c2cd" [ 926.086019] env[62183]: _type = "Task" [ 926.086019] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.093527] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5248eb77-b982-4403-ab36-9499b422c2cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.123058] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a18d34-9fa5-4a67-a55d-7838b484e5eb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.130985] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93639f00-95d5-4f24-b2ab-b95b94786640 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.137060] env[62183]: DEBUG oslo_concurrency.lockutils [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.165814] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46e93d3-02b5-4c2f-b36d-055eca85c550 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.176137] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049ea6ac-845e-43b3-aae4-f0d315ffe0ea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.190441] env[62183]: DEBUG nova.compute.provider_tree [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.209956] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387231, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.370034] env[62183]: DEBUG nova.network.neutron [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Updated VIF entry in instance network info cache for port ef8e8a41-f9ce-44f2-bcea-feef82484547. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 926.370442] env[62183]: DEBUG nova.network.neutron [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Updating instance_info_cache with network_info: [{"id": "ef8e8a41-f9ce-44f2-bcea-feef82484547", "address": "fa:16:3e:db:a5:ae", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef8e8a41-f9", "ovs_interfaceid": "ef8e8a41-f9ce-44f2-bcea-feef82484547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.447382] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.447648] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.447858] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.448058] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.448235] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.450313] env[62183]: INFO nova.compute.manager [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Terminating instance [ 926.451981] env[62183]: DEBUG nova.compute.manager [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.452193] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.453032] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76939652-9157-4b2e-9949-74a4fad50b83 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.460099] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.460400] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee954116-6d27-4b32-970c-a9149e748a10 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.465705] env[62183]: DEBUG oslo_vmware.api [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 926.465705] env[62183]: value = "task-1387235" [ 926.465705] env[62183]: _type = "Task" [ 926.465705] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.474039] env[62183]: DEBUG oslo_vmware.api [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387235, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.551896] env[62183]: DEBUG oslo_vmware.api [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387233, 'name': PowerOnVM_Task, 'duration_secs': 0.846999} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.552199] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.552440] env[62183]: INFO nova.compute.manager [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Took 8.45 seconds to spawn the instance on the hypervisor. [ 926.552649] env[62183]: DEBUG nova.compute.manager [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.553630] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1692bb40-2e37-48e7-97c9-ead5bdc33f00 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.595216] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5248eb77-b982-4403-ab36-9499b422c2cd, 'name': SearchDatastore_Task, 'duration_secs': 0.008495} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.595990] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d380f62-cb31-40f9-bb14-d91bc396cff6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.601698] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 926.601698] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52af5d64-5f8e-9d73-1f33-704e0fd3dd2d" [ 926.601698] env[62183]: _type = "Task" [ 926.601698] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.611725] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52af5d64-5f8e-9d73-1f33-704e0fd3dd2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.693375] env[62183]: DEBUG nova.scheduler.client.report [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 926.713557] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387231, 'name': CloneVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.872836] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Releasing lock "refresh_cache-7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.873185] env[62183]: DEBUG nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received event network-vif-unplugged-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.873407] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Acquiring lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.873617] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.873781] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.873949] env[62183]: DEBUG nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] No waiting events found dispatching network-vif-unplugged-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 926.874139] env[62183]: WARNING nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received unexpected event network-vif-unplugged-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 for instance with vm_state shelved and task_state shelving_offloading. [ 926.874341] env[62183]: DEBUG nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received event network-changed-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.874502] env[62183]: DEBUG nova.compute.manager [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Refreshing instance network info cache due to event network-changed-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 926.874701] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Acquiring lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.874839] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Acquired lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.875009] env[62183]: DEBUG nova.network.neutron [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Refreshing network info cache for port 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 926.975315] env[62183]: DEBUG oslo_vmware.api [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387235, 'name': PowerOffVM_Task, 'duration_secs': 0.184971} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.975620] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.975794] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.976058] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a32ac2c3-7266-4cae-b19c-83d10ff20ee0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.071745] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6111bc-759b-48ff-a3c0-6a49d9cd96fa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.077061] env[62183]: INFO nova.compute.manager [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Took 13.28 seconds to build instance. [ 927.092085] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance '28d279bd-206e-4ba0-bf84-2980ab41f38d' progress to 0 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 927.111983] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52af5d64-5f8e-9d73-1f33-704e0fd3dd2d, 'name': SearchDatastore_Task, 'duration_secs': 0.030682} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.112320] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.112627] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631/7b8a64ad-afcb-4e7c-8c00-86f69d8d3631.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 927.112922] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b67254f-8f7c-44a4-985f-8a22b822dc38 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.119625] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 927.119625] env[62183]: value = "task-1387237" [ 927.119625] env[62183]: _type = "Task" [ 927.119625] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.128428] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.154047] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.154143] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.154338] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Deleting the datastore file [datastore1] ac555ffc-ce4e-4650-97fd-c26a3246fe4b {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.154796] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e10908e-dcc2-4a30-b345-525ef99cbaac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.161943] env[62183]: DEBUG oslo_vmware.api [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for the task: (returnval){ [ 927.161943] env[62183]: value = "task-1387238" [ 927.161943] env[62183]: _type = "Task" [ 927.161943] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.170774] env[62183]: DEBUG oslo_vmware.api [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387238, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.198404] env[62183]: DEBUG oslo_concurrency.lockutils [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.736s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.201019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.064s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.201949] env[62183]: DEBUG nova.objects.instance [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lazy-loading 'resources' on Instance uuid fb348784-62a0-4d1f-ac7f-f176f3da0dd9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.211671] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387231, 'name': CloneVM_Task, 'duration_secs': 1.53405} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.212687] env[62183]: INFO nova.virt.vmwareapi.vmops [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Created linked-clone VM from snapshot [ 927.215676] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedffac1-cf6c-464d-8784-a2541970e2c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.221718] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Uploading image e82a77f9-6e38-4522-9049-20bc607f6a15 {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 927.224610] env[62183]: INFO nova.scheduler.client.report [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Deleted allocations for instance 46aa1062-3fcd-4581-8647-32e1372e9491 [ 927.248254] env[62183]: DEBUG oslo_vmware.rw_handles [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 927.248254] env[62183]: value = "vm-294484" [ 927.248254] env[62183]: _type = "VirtualMachine" [ 927.248254] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 927.248254] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-be363f63-fd39-4d7d-aa20-91e782c81f01 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.255898] env[62183]: DEBUG oslo_vmware.rw_handles [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease: (returnval){ [ 927.255898] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52175f78-9157-850a-2ed9-c03df47fce21" [ 927.255898] env[62183]: _type = "HttpNfcLease" [ 927.255898] env[62183]: } obtained for exporting VM: (result){ [ 927.255898] env[62183]: value = "vm-294484" [ 927.255898] env[62183]: _type = "VirtualMachine" [ 927.255898] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 927.256776] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the lease: (returnval){ [ 927.256776] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52175f78-9157-850a-2ed9-c03df47fce21" [ 927.256776] env[62183]: _type = "HttpNfcLease" [ 927.256776] env[62183]: } to be ready. {{(pid=62183) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 927.264580] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 927.264580] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52175f78-9157-850a-2ed9-c03df47fce21" [ 927.264580] env[62183]: _type = "HttpNfcLease" [ 927.264580] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 927.580360] env[62183]: DEBUG oslo_concurrency.lockutils [None req-480d2a62-6830-4776-b37e-0a353e909edf tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.789s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.601084] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.601473] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24834db8-7b07-4d1e-8610-fcc7d2513c69 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.608971] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 927.608971] env[62183]: value = "task-1387240" [ 927.608971] env[62183]: _type = "Task" [ 927.608971] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.623199] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.637253] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387237, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505423} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.637781] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631/7b8a64ad-afcb-4e7c-8c00-86f69d8d3631.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 927.637999] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 927.638274] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08af1dc8-e5e0-4b26-b976-ddbc8de6a00c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.645756] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 927.645756] env[62183]: value = "task-1387241" [ 927.645756] env[62183]: _type = "Task" [ 927.645756] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.653714] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.672115] env[62183]: DEBUG oslo_vmware.api [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Task: {'id': task-1387238, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.445923} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.672406] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.672593] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.672807] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.672990] env[62183]: INFO nova.compute.manager [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Took 1.22 seconds to destroy the instance on the hypervisor. [ 927.673275] env[62183]: DEBUG oslo.service.loopingcall [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.673595] env[62183]: DEBUG nova.compute.manager [-] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.673697] env[62183]: DEBUG nova.network.neutron [-] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.704095] env[62183]: DEBUG nova.objects.instance [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lazy-loading 'numa_topology' on Instance uuid fb348784-62a0-4d1f-ac7f-f176f3da0dd9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.732655] env[62183]: DEBUG oslo_concurrency.lockutils [None req-314017ee-f823-4a25-8125-42df44eec6b6 tempest-ImagesTestJSON-1973005259 tempest-ImagesTestJSON-1973005259-project-member] Lock "46aa1062-3fcd-4581-8647-32e1372e9491" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.655s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.765291] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 927.765291] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52175f78-9157-850a-2ed9-c03df47fce21" [ 927.765291] env[62183]: _type = "HttpNfcLease" [ 927.765291] env[62183]: } is ready. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 927.765617] env[62183]: DEBUG oslo_vmware.rw_handles [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 927.765617] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52175f78-9157-850a-2ed9-c03df47fce21" [ 927.765617] env[62183]: _type = "HttpNfcLease" [ 927.765617] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 927.766348] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5bad72-2577-40c5-8bf1-2c1cfeda3cbe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.773264] env[62183]: DEBUG oslo_vmware.rw_handles [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528c605a-f842-6f09-1205-4ecbe6e0c821/disk-0.vmdk from lease info. {{(pid=62183) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 927.773458] env[62183]: DEBUG oslo_vmware.rw_handles [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528c605a-f842-6f09-1205-4ecbe6e0c821/disk-0.vmdk for reading. {{(pid=62183) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 927.831746] env[62183]: DEBUG nova.network.neutron [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updated VIF entry in instance network info cache for port 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.832742] env[62183]: DEBUG nova.network.neutron [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updating instance_info_cache with network_info: [{"id": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "address": "fa:16:3e:3f:c5:6f", "network": {"id": "7d52241a-8c3c-46bc-b64d-1cbc74ec2434", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-104877467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e74d26cb8f5a47a48dc025b8058c5768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap37fafdd3-76", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.907422] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f6b43284-2941-4af1-8127-e97beef11325 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.975773] env[62183]: DEBUG nova.compute.manager [req-98048bc3-0737-4821-b020-0a18f4791f91 req-91f2d176-6465-4c0a-a724-318d0d970dc2 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Received event network-vif-deleted-91e1877d-c4f8-4ad7-8076-08c985c66e4c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.976450] env[62183]: INFO nova.compute.manager [req-98048bc3-0737-4821-b020-0a18f4791f91 req-91f2d176-6465-4c0a-a724-318d0d970dc2 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Neutron deleted interface 91e1877d-c4f8-4ad7-8076-08c985c66e4c; detaching it from the instance and deleting it from the info cache [ 927.977044] env[62183]: DEBUG nova.network.neutron [req-98048bc3-0737-4821-b020-0a18f4791f91 req-91f2d176-6465-4c0a-a724-318d0d970dc2 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.119915] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387240, 'name': PowerOffVM_Task, 'duration_secs': 0.346448} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.120224] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 928.120418] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance '28d279bd-206e-4ba0-bf84-2980ab41f38d' progress to 17 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 928.142385] env[62183]: DEBUG nova.compute.manager [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Stashing vm_state: active {{(pid=62183) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 928.157315] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.158945} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.157606] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.158539] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a9ee5a-3744-40ff-93ae-954dba6f7727 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.180048] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631/7b8a64ad-afcb-4e7c-8c00-86f69d8d3631.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.181092] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0447954-8da9-4676-8879-cd71b20db9d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.201579] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 928.201579] env[62183]: value = "task-1387242" [ 928.201579] env[62183]: _type = "Task" [ 928.201579] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.206843] env[62183]: DEBUG nova.objects.base [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 928.212275] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387242, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.232012] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Volume attach. Driver type: vmdk {{(pid=62183) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 928.232289] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294482', 'volume_id': 'f481c3e9-7f36-4566-9973-b98f2c584ee9', 'name': 'volume-f481c3e9-7f36-4566-9973-b98f2c584ee9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de', 'attached_at': '', 'detached_at': '', 'volume_id': 'f481c3e9-7f36-4566-9973-b98f2c584ee9', 'serial': 'f481c3e9-7f36-4566-9973-b98f2c584ee9'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 928.233739] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3395aeea-99f6-4079-b045-ed04bc2b134d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.253568] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3da1c1-2fab-4b63-b96f-acb094013f6b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.280631] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] volume-f481c3e9-7f36-4566-9973-b98f2c584ee9/volume-f481c3e9-7f36-4566-9973-b98f2c584ee9.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.283755] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9fd440e-8a14-41ca-8367-85fa6b829d91 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.301617] env[62183]: DEBUG oslo_vmware.api [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 928.301617] env[62183]: value = "task-1387243" [ 928.301617] env[62183]: _type = "Task" [ 928.301617] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.309874] env[62183]: DEBUG oslo_vmware.api [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387243, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.335669] env[62183]: DEBUG oslo_concurrency.lockutils [req-74069f55-de5e-447e-b928-80fc273460b0 req-b1b56488-bf51-410a-ba86-4047da21f3c6 service nova] Releasing lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.413599] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cbd684-26f7-411b-aa0d-6ff4de1e1db6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.420800] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72a803a-e5ef-41ab-9d68-447d6dc6f6bf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.453775] env[62183]: DEBUG nova.network.neutron [-] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.455974] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083f1410-3c54-4b53-95d2-a62ffda4b872 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.464774] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0578ae-6ce4-413c-8567-9b5c5315d8e0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.481037] env[62183]: DEBUG nova.compute.provider_tree [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.482225] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c6a72d7-d1f7-43d6-89fb-2d29773e4498 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.490622] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea815a2d-7d24-4876-812c-cd0f7b2cf5ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.503711] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.521097] env[62183]: DEBUG nova.compute.manager [req-98048bc3-0737-4821-b020-0a18f4791f91 req-91f2d176-6465-4c0a-a724-318d0d970dc2 service nova] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Detach interface failed, port_id=91e1877d-c4f8-4ad7-8076-08c985c66e4c, reason: Instance ac555ffc-ce4e-4650-97fd-c26a3246fe4b could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 928.626988] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.627431] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.627692] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.628734] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.628734] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.628734] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.628734] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.628734] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.628734] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.629019] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.629195] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.634491] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c970cc2-c107-42de-b764-5db888b87e9d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.654213] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 928.654213] env[62183]: value = "task-1387244" [ 928.654213] env[62183]: _type = "Task" [ 928.654213] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.662827] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387244, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.663923] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.711259] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387242, 'name': ReconfigVM_Task, 'duration_secs': 0.291709} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.712024] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631/7b8a64ad-afcb-4e7c-8c00-86f69d8d3631.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.712579] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a5e9694-f687-4ca1-abbd-daf3021e96b6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.719151] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 928.719151] env[62183]: value = "task-1387245" [ 928.719151] env[62183]: _type = "Task" [ 928.719151] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.728032] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387245, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.813523] env[62183]: DEBUG oslo_vmware.api [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387243, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.961093] env[62183]: INFO nova.compute.manager [-] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Took 1.29 seconds to deallocate network for instance. [ 928.986218] env[62183]: DEBUG nova.scheduler.client.report [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 929.165805] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387244, 'name': ReconfigVM_Task, 'duration_secs': 0.151477} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.166313] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance '28d279bd-206e-4ba0-bf84-2980ab41f38d' progress to 33 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 929.231476] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387245, 'name': Rename_Task, 'duration_secs': 0.163443} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.231802] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 929.232074] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa25a89d-ecbe-4b69-8ee0-a328a9f50c5d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.239759] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 929.239759] env[62183]: value = "task-1387246" [ 929.239759] env[62183]: _type = "Task" [ 929.239759] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.251930] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387246, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.313451] env[62183]: DEBUG oslo_vmware.api [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387243, 'name': ReconfigVM_Task, 'duration_secs': 0.607839} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.313745] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Reconfigured VM instance instance-0000004f to attach disk [datastore1] volume-f481c3e9-7f36-4566-9973-b98f2c584ee9/volume-f481c3e9-7f36-4566-9973-b98f2c584ee9.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.319394] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d823ad2-1e50-4689-8ae8-9751ddce7466 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.334324] env[62183]: DEBUG oslo_vmware.api [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 929.334324] env[62183]: value = "task-1387247" [ 929.334324] env[62183]: _type = "Task" [ 929.334324] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.342753] env[62183]: DEBUG oslo_vmware.api [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387247, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.471171] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.495650] env[62183]: DEBUG oslo_concurrency.lockutils [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.295s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.498901] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.835s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.674658] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.675075] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.675247] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.675551] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.675800] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.676770] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.676770] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.676893] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.677129] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.677203] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.678309] env[62183]: DEBUG nova.virt.hardware [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.683662] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Reconfiguring VM instance instance-00000054 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 929.684021] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b362f4c1-ec25-4c88-983e-a18e24a6bcdd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.703314] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 929.703314] env[62183]: value = "task-1387248" [ 929.703314] env[62183]: _type = "Task" [ 929.703314] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.712300] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387248, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.750259] env[62183]: DEBUG oslo_vmware.api [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387246, 'name': PowerOnVM_Task, 'duration_secs': 0.453711} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.750604] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.750881] env[62183]: INFO nova.compute.manager [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Took 6.27 seconds to spawn the instance on the hypervisor. [ 929.751084] env[62183]: DEBUG nova.compute.manager [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.751914] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c427feee-d0c8-49ff-8121-37815a75844c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.844441] env[62183]: DEBUG oslo_vmware.api [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387247, 'name': ReconfigVM_Task, 'duration_secs': 0.149226} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.844783] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294482', 'volume_id': 'f481c3e9-7f36-4566-9973-b98f2c584ee9', 'name': 'volume-f481c3e9-7f36-4566-9973-b98f2c584ee9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de', 'attached_at': '', 'detached_at': '', 'volume_id': 'f481c3e9-7f36-4566-9973-b98f2c584ee9', 'serial': 'f481c3e9-7f36-4566-9973-b98f2c584ee9'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 930.008728] env[62183]: INFO nova.compute.claims [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.019860] env[62183]: DEBUG oslo_concurrency.lockutils [None req-872a5ae0-a840-47a0-b227-ae22fdcd1e7f tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.165s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.020865] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.517s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.023485] env[62183]: INFO nova.compute.manager [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Unshelving [ 930.214144] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387248, 'name': ReconfigVM_Task, 'duration_secs': 0.236046} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.214456] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Reconfigured VM instance instance-00000054 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 930.215341] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7058e1e0-c654-4da1-b674-cb9df66cb0f8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.237658] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 28d279bd-206e-4ba0-bf84-2980ab41f38d/28d279bd-206e-4ba0-bf84-2980ab41f38d.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.237984] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f87ec06-51aa-4685-9388-44317aa44978 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.256252] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 930.256252] env[62183]: value = "task-1387249" [ 930.256252] env[62183]: _type = "Task" [ 930.256252] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.270089] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387249, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.272449] env[62183]: INFO nova.compute.manager [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Took 11.17 seconds to build instance. [ 930.521728] env[62183]: INFO nova.compute.resource_tracker [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating resource usage from migration bd8cf8f4-2aae-4118-8c5a-2cdbfb615009 [ 930.681796] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11f2060-77d8-4d60-8316-c089d67e644b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.700419] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b0304d-e8ff-4f47-a9fb-3ae060cdff0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.728334] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ef55f7-1574-4fd7-aa91-d3dc41ea364f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.735865] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3621439e-822b-4a87-a59a-0321d6e36fcf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.749516] env[62183]: DEBUG nova.compute.provider_tree [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 930.766549] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387249, 'name': ReconfigVM_Task, 'duration_secs': 0.448318} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.766890] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 28d279bd-206e-4ba0-bf84-2980ab41f38d/28d279bd-206e-4ba0-bf84-2980ab41f38d.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.767223] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance '28d279bd-206e-4ba0-bf84-2980ab41f38d' progress to 50 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 930.774215] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1ba0a2b2-7b41-44e2-9e82-f690742dec17 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.677s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.891757] env[62183]: DEBUG nova.objects.instance [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'flavor' on Instance uuid ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.058102] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.274068] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6dc28bc-cd83-4526-9f3c-802630dc5d52 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.280877] env[62183]: ERROR nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [req-5d0a9f80-3bbf-40ef-a940-2fd34f0808f2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5d0a9f80-3bbf-40ef-a940-2fd34f0808f2"}]} [ 931.304344] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b105c77-5137-431f-b275-ef9554255d12 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.310321] env[62183]: DEBUG nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 931.334036] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance '28d279bd-206e-4ba0-bf84-2980ab41f38d' progress to 67 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 931.336684] env[62183]: DEBUG nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 931.336969] env[62183]: DEBUG nova.compute.provider_tree [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 931.349464] env[62183]: DEBUG nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 931.378022] env[62183]: DEBUG nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 931.398216] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f7ef7b0b-ac9c-481a-b870-ffff554aaf89 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.788s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.574054] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49311cc2-fc5c-4bc2-a3c4-967c586af910 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.588695] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b1b39c-e361-4d84-959e-c24d482726b4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.622204] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af00c00-dbf4-4440-a4c6-dd9315e50b0f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.630426] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfd1c63-78c6-41fc-9292-9dac15b6c35f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.645229] env[62183]: DEBUG nova.compute.provider_tree [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 931.873263] env[62183]: DEBUG nova.network.neutron [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Port f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc binding to destination host cpu-1 is already ACTIVE {{(pid=62183) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 931.946500] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.946804] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.947032] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.947239] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.947421] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.950298] env[62183]: INFO nova.compute.manager [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Terminating instance [ 931.952125] env[62183]: DEBUG nova.compute.manager [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 931.952327] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.953152] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f9adb1-8f19-4eb7-93db-c9813dabb26d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.961306] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.961548] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20420d02-a70e-44c7-b036-50c8e7b9c882 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.968019] env[62183]: DEBUG oslo_vmware.api [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 931.968019] env[62183]: value = "task-1387250" [ 931.968019] env[62183]: _type = "Task" [ 931.968019] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.975831] env[62183]: DEBUG oslo_vmware.api [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387250, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.165823] env[62183]: ERROR nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [req-7243484a-754d-4629-a862-e28a235d152a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7243484a-754d-4629-a862-e28a235d152a"}]} [ 932.182298] env[62183]: DEBUG nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 932.195964] env[62183]: DEBUG nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 932.196599] env[62183]: DEBUG nova.compute.provider_tree [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.209734] env[62183]: DEBUG nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 932.227891] env[62183]: DEBUG nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 932.370120] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.370781] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.408155] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629e4e19-12ef-4cd6-a849-3c85d7a37974 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.416209] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b69e945-28dc-4b8b-9917-713620b11296 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.449674] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e54f4d2-f3c2-4192-accf-7b75c41b659c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.457717] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ef85d2-d6e2-4525-a6f4-9e25fe347d01 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.472448] env[62183]: DEBUG nova.compute.provider_tree [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.481608] env[62183]: DEBUG oslo_vmware.api [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387250, 'name': PowerOffVM_Task, 'duration_secs': 0.195007} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.482429] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 932.482609] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 932.482864] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a86a4dd5-5e17-406e-9a10-827f794f23bd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.562552] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 932.562793] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 932.562981] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleting the datastore file [datastore1] 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.563267] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fb6a1a7-f2a5-43ec-8d93-2572a685c6bf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.570093] env[62183]: DEBUG oslo_vmware.api [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 932.570093] env[62183]: value = "task-1387252" [ 932.570093] env[62183]: _type = "Task" [ 932.570093] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.578078] env[62183]: DEBUG oslo_vmware.api [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387252, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.873821] env[62183]: DEBUG nova.compute.manager [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.896671] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "28d279bd-206e-4ba0-bf84-2980ab41f38d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.896980] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.897224] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.015613] env[62183]: DEBUG nova.scheduler.client.report [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 105 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 933.015947] env[62183]: DEBUG nova.compute.provider_tree [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 105 to 106 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 933.016159] env[62183]: DEBUG nova.compute.provider_tree [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.080551] env[62183]: DEBUG oslo_vmware.api [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387252, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.315189} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.081046] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.081145] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.081246] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.081425] env[62183]: INFO nova.compute.manager [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Took 1.13 seconds to destroy the instance on the hypervisor. [ 933.081867] env[62183]: DEBUG oslo.service.loopingcall [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.082152] env[62183]: DEBUG nova.compute.manager [-] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 933.082259] env[62183]: DEBUG nova.network.neutron [-] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 933.407668] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.523367] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 4.024s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.523599] env[62183]: INFO nova.compute.manager [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Migrating [ 933.538621] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.067s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.539428] env[62183]: DEBUG nova.objects.instance [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lazy-loading 'resources' on Instance uuid ac555ffc-ce4e-4650-97fd-c26a3246fe4b {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.551387] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.551910] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.551910] env[62183]: INFO nova.compute.manager [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Rebooting instance [ 933.703278] env[62183]: DEBUG nova.compute.manager [req-b84abc54-8b34-4800-9c9b-e63f9ae3b793 req-38256623-0080-489a-af80-63af4a94b49b service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Received event network-vif-deleted-ef8e8a41-f9ce-44f2-bcea-feef82484547 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.703524] env[62183]: INFO nova.compute.manager [req-b84abc54-8b34-4800-9c9b-e63f9ae3b793 req-38256623-0080-489a-af80-63af4a94b49b service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Neutron deleted interface ef8e8a41-f9ce-44f2-bcea-feef82484547; detaching it from the instance and deleting it from the info cache [ 933.703782] env[62183]: DEBUG nova.network.neutron [req-b84abc54-8b34-4800-9c9b-e63f9ae3b793 req-38256623-0080-489a-af80-63af4a94b49b service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.975372] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.975645] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.975876] env[62183]: DEBUG nova.network.neutron [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.046870] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.047695] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.048077] env[62183]: DEBUG nova.network.neutron [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.081960] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.081960] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.082119] env[62183]: DEBUG nova.network.neutron [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.173973] env[62183]: DEBUG nova.network.neutron [-] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.206420] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d626e025-3d87-4a96-b468-5a1a8c7169f9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.218398] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96770c9-38d5-4eb5-961c-456c8712a8ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.247748] env[62183]: DEBUG nova.compute.manager [req-b84abc54-8b34-4800-9c9b-e63f9ae3b793 req-38256623-0080-489a-af80-63af4a94b49b service nova] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Detach interface failed, port_id=ef8e8a41-f9ce-44f2-bcea-feef82484547, reason: Instance 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 934.261330] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcaa4cec-6fa8-4ff5-b206-d557a49c2673 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.268931] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dcd71ff-5fdd-48a7-88d7-be85f6ca35b3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.299265] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3126ec2-72e4-4669-94a0-92d91dbaddf8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.306853] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022cba3d-6247-4744-afc5-9bdc9379f2ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.320166] env[62183]: DEBUG nova.compute.provider_tree [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.676741] env[62183]: INFO nova.compute.manager [-] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Took 1.59 seconds to deallocate network for instance. [ 934.827304] env[62183]: DEBUG nova.scheduler.client.report [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 934.850202] env[62183]: DEBUG nova.network.neutron [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance_info_cache with network_info: [{"id": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "address": "fa:16:3e:39:9c:d2", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2ec12e2-ad", "ovs_interfaceid": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.877846] env[62183]: DEBUG nova.network.neutron [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance_info_cache with network_info: [{"id": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "address": "fa:16:3e:a4:03:ac", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1af05d2-92", "ovs_interfaceid": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.933069] env[62183]: DEBUG nova.network.neutron [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.184055] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.329862] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.332209] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.274s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.332437] env[62183]: DEBUG nova.objects.instance [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lazy-loading 'pci_requests' on Instance uuid fb348784-62a0-4d1f-ac7f-f176f3da0dd9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.348432] env[62183]: INFO nova.scheduler.client.report [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Deleted allocations for instance ac555ffc-ce4e-4650-97fd-c26a3246fe4b [ 935.353137] env[62183]: DEBUG oslo_concurrency.lockutils [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.381534] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.435836] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.438370] env[62183]: DEBUG nova.compute.manager [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.439793] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed7b27f-be34-40e6-9fe2-3d6ce2c9c840 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.592612] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.593030] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.626066] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "d30cfb69-1814-49b2-a29d-10cfbcc05722" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.626296] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "d30cfb69-1814-49b2-a29d-10cfbcc05722" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.837256] env[62183]: DEBUG nova.objects.instance [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lazy-loading 'numa_topology' on Instance uuid fb348784-62a0-4d1f-ac7f-f176f3da0dd9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.856885] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af7c006a-49f7-4cb1-a385-696a9014fae6 tempest-AttachInterfacesTestJSON-544859338 tempest-AttachInterfacesTestJSON-544859338-project-member] Lock "ac555ffc-ce4e-4650-97fd-c26a3246fe4b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.409s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.876917] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d7c5cd-31a6-41b3-b1f4-d488d4a92b88 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.899737] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b774b1-3919-4227-a2fa-c548734d5aca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.906932] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance '28d279bd-206e-4ba0-bf84-2980ab41f38d' progress to 83 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 936.095368] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 936.128914] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 936.342606] env[62183]: INFO nova.compute.claims [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.413283] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.413920] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e28c7487-f7fe-4029-818c-5aad01845486 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.421332] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 936.421332] env[62183]: value = "task-1387253" [ 936.421332] env[62183]: _type = "Task" [ 936.421332] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.430945] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387253, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.455802] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc109eb-0bea-424a-87ef-926d18924e83 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.463205] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Doing hard reboot of VM {{(pid=62183) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 936.464703] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-29e17efc-a2fb-471b-b79d-b45eeae57509 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.469871] env[62183]: DEBUG oslo_vmware.api [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 936.469871] env[62183]: value = "task-1387254" [ 936.469871] env[62183]: _type = "Task" [ 936.469871] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.481277] env[62183]: DEBUG oslo_vmware.api [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387254, 'name': ResetVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.624265] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.646796] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.908962] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5ad5d2-a6a6-44a2-a109-8930b15b4bca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.927921] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance 'b89455d9-6d00-41ac-95ec-fb036b8c26c9' progress to 0 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 936.942022] env[62183]: DEBUG oslo_vmware.api [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387253, 'name': PowerOnVM_Task, 'duration_secs': 0.473008} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.942022] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 936.942022] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-191f5baf-f847-4bf8-8c29-410fe201f8f5 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance '28d279bd-206e-4ba0-bf84-2980ab41f38d' progress to 100 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 936.979952] env[62183]: DEBUG oslo_vmware.api [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387254, 'name': ResetVM_Task, 'duration_secs': 0.094592} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.981361] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Did hard reboot of VM {{(pid=62183) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 936.981564] env[62183]: DEBUG nova.compute.manager [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.982530] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab13d8a3-dc84-4e53-8724-df365e1c1d95 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.437162] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.437474] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db450bf6-c74f-4c1d-a923-444a57e2b0b1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.444745] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 937.444745] env[62183]: value = "task-1387255" [ 937.444745] env[62183]: _type = "Task" [ 937.444745] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.458020] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387255, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.496468] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c53fb78d-4dbf-4ed1-9493-bb10861e5960 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.945s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.547872] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c3b75d-0a75-4364-8a76-50eb99291dec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.556741] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f2be7b-02f7-4b1b-bdfb-2be914bf2b23 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.594250] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0029a61b-14b5-45e0-8e91-a3f12e088093 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.604950] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14180197-9523-40c6-896f-5b8376c82b58 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.623696] env[62183]: DEBUG nova.compute.provider_tree [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.662820] env[62183]: DEBUG oslo_vmware.rw_handles [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528c605a-f842-6f09-1205-4ecbe6e0c821/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 937.663866] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10171b89-ec38-4452-a855-fe196c1e7f24 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.671154] env[62183]: DEBUG oslo_vmware.rw_handles [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528c605a-f842-6f09-1205-4ecbe6e0c821/disk-0.vmdk is in state: ready. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 937.671338] env[62183]: ERROR oslo_vmware.rw_handles [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528c605a-f842-6f09-1205-4ecbe6e0c821/disk-0.vmdk due to incomplete transfer. [ 937.671615] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5f54c74b-1c99-4c36-a855-0e19b5c435dd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.678641] env[62183]: DEBUG oslo_vmware.rw_handles [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528c605a-f842-6f09-1205-4ecbe6e0c821/disk-0.vmdk. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 937.678851] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Uploaded image e82a77f9-6e38-4522-9049-20bc607f6a15 to the Glance image server {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 937.681372] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Destroying the VM {{(pid=62183) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 937.681644] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-12f80b33-57cb-4ff0-b27e-3b7ac43f34c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.688190] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 937.688190] env[62183]: value = "task-1387256" [ 937.688190] env[62183]: _type = "Task" [ 937.688190] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.697789] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387256, 'name': Destroy_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.960709] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387255, 'name': PowerOffVM_Task, 'duration_secs': 0.384857} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.960709] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.960709] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance 'b89455d9-6d00-41ac-95ec-fb036b8c26c9' progress to 17 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 938.130226] env[62183]: DEBUG nova.scheduler.client.report [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 938.198298] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387256, 'name': Destroy_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.465979] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.466351] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.466560] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.466838] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.467045] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.467279] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.467582] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.467830] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.468067] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.468346] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.468654] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.475893] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1db4a9fb-6f22-4fc9-89e0-9b2167f136f8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.494540] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 938.494540] env[62183]: value = "task-1387257" [ 938.494540] env[62183]: _type = "Task" [ 938.494540] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.505680] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387257, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.635148] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.303s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.637532] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.230s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.639194] env[62183]: INFO nova.compute.claims [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 938.699314] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387256, 'name': Destroy_Task, 'duration_secs': 0.755473} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.699631] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Destroyed the VM [ 938.699908] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deleting Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 938.700197] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0656e019-64a3-4a6b-a0b7-8614d16d745b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.707622] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 938.707622] env[62183]: value = "task-1387258" [ 938.707622] env[62183]: _type = "Task" [ 938.707622] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.717204] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387258, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.886221] env[62183]: INFO nova.network.neutron [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updating port 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 939.006933] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387257, 'name': ReconfigVM_Task, 'duration_secs': 0.337309} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.006933] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance 'b89455d9-6d00-41ac-95ec-fb036b8c26c9' progress to 33 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 939.217107] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387258, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.523051] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.523485] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.523694] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.524054] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.524269] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.524468] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.524739] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.525076] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.525383] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.525675] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.526040] env[62183]: DEBUG nova.virt.hardware [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.532230] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Reconfiguring VM instance instance-00000055 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 939.532568] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33cc3a73-5959-4c6d-9b32-c28e3e4ce02c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.553094] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 939.553094] env[62183]: value = "task-1387259" [ 939.553094] env[62183]: _type = "Task" [ 939.553094] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.561026] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387259, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.638090] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "28d279bd-206e-4ba0-bf84-2980ab41f38d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.638400] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.638636] env[62183]: DEBUG nova.compute.manager [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Going to confirm migration 2 {{(pid=62183) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 939.718279] env[62183]: DEBUG oslo_vmware.api [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387258, 'name': RemoveSnapshot_Task, 'duration_secs': 0.906245} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.718564] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deleted Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 939.718811] env[62183]: INFO nova.compute.manager [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Took 17.10 seconds to snapshot the instance on the hypervisor. [ 939.842390] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2bdfe3-9362-432b-bf4f-8147f4dd73ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.849751] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682b03d8-2f3d-4442-b307-61670f8e8c7d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.881306] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6109a9-2ec9-45c3-8da7-60685d946748 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.889431] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ec98cd-062a-454b-b9c8-090e00560981 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.904142] env[62183]: DEBUG nova.compute.provider_tree [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.067060] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387259, 'name': ReconfigVM_Task, 'duration_secs': 0.417486} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.067060] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Reconfigured VM instance instance-00000055 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 940.067060] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50758604-eb60-40e2-90b7-1c1e07074d0b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.090064] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] b89455d9-6d00-41ac-95ec-fb036b8c26c9/b89455d9-6d00-41ac-95ec-fb036b8c26c9.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.090569] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2ef2fe3-6402-4500-8d68-aca5e660dbf3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.110348] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 940.110348] env[62183]: value = "task-1387260" [ 940.110348] env[62183]: _type = "Task" [ 940.110348] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.118764] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387260, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.258987] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.258987] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.258987] env[62183]: DEBUG nova.network.neutron [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.258987] env[62183]: DEBUG nova.objects.instance [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lazy-loading 'info_cache' on Instance uuid 28d279bd-206e-4ba0-bf84-2980ab41f38d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.310688] env[62183]: DEBUG nova.compute.manager [None req-9339d303-6d4e-48b8-a1e5-fcf72c771de8 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Found 2 images (rotation: 2) {{(pid=62183) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 940.408660] env[62183]: DEBUG nova.scheduler.client.report [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 940.419055] env[62183]: DEBUG nova.compute.manager [req-1bc37ac2-0ad8-47a7-b61e-7861928c50c9 req-a8819438-f113-46a1-873b-12a00262778b service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received event network-vif-plugged-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.419055] env[62183]: DEBUG oslo_concurrency.lockutils [req-1bc37ac2-0ad8-47a7-b61e-7861928c50c9 req-a8819438-f113-46a1-873b-12a00262778b service nova] Acquiring lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.419055] env[62183]: DEBUG oslo_concurrency.lockutils [req-1bc37ac2-0ad8-47a7-b61e-7861928c50c9 req-a8819438-f113-46a1-873b-12a00262778b service nova] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.419055] env[62183]: DEBUG oslo_concurrency.lockutils [req-1bc37ac2-0ad8-47a7-b61e-7861928c50c9 req-a8819438-f113-46a1-873b-12a00262778b service nova] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.419055] env[62183]: DEBUG nova.compute.manager [req-1bc37ac2-0ad8-47a7-b61e-7861928c50c9 req-a8819438-f113-46a1-873b-12a00262778b service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] No waiting events found dispatching network-vif-plugged-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 940.419055] env[62183]: WARNING nova.compute.manager [req-1bc37ac2-0ad8-47a7-b61e-7861928c50c9 req-a8819438-f113-46a1-873b-12a00262778b service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received unexpected event network-vif-plugged-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 for instance with vm_state shelved_offloaded and task_state spawning. [ 940.470646] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.470646] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.470646] env[62183]: DEBUG nova.network.neutron [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.623794] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387260, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.919922] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.920548] env[62183]: DEBUG nova.compute.manager [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 940.923325] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.739s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.923572] env[62183]: DEBUG nova.objects.instance [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lazy-loading 'resources' on Instance uuid 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.121413] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387260, 'name': ReconfigVM_Task, 'duration_secs': 0.715229} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.123973] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Reconfigured VM instance instance-00000055 to attach disk [datastore1] b89455d9-6d00-41ac-95ec-fb036b8c26c9/b89455d9-6d00-41ac-95ec-fb036b8c26c9.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.123973] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance 'b89455d9-6d00-41ac-95ec-fb036b8c26c9' progress to 50 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 941.308190] env[62183]: DEBUG nova.network.neutron [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updating instance_info_cache with network_info: [{"id": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "address": "fa:16:3e:3f:c5:6f", "network": {"id": "7d52241a-8c3c-46bc-b64d-1cbc74ec2434", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-104877467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e74d26cb8f5a47a48dc025b8058c5768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fafdd3-76", "ovs_interfaceid": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.426461] env[62183]: DEBUG nova.compute.utils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 941.431270] env[62183]: DEBUG nova.compute.manager [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 941.431270] env[62183]: DEBUG nova.network.neutron [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 941.499228] env[62183]: DEBUG nova.policy [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ae1837d5db145278417f7cdd55a3fea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06e81bd1a81d4009ae2a75fe819f9b7c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 941.636346] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd64c89-2072-46e3-82eb-95539f8c51b8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.660902] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023e3361-0ba9-4022-9e48-623ff072c12d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.663887] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dad8c22-4515-4b32-82a2-f291a502bb19 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.684145] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca60316-b9c9-45a9-b619-8db8eb16c552 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.687273] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance 'b89455d9-6d00-41ac-95ec-fb036b8c26c9' progress to 67 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 941.719518] env[62183]: DEBUG nova.network.neutron [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance_info_cache with network_info: [{"id": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "address": "fa:16:3e:39:9c:d2", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2ec12e2-ad", "ovs_interfaceid": "f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.720965] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee6a334-4187-4543-8856-19c06a2512dd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.732850] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9570cf05-c680-4d1f-b17d-1fbe9951c74c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.746960] env[62183]: DEBUG nova.compute.provider_tree [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.813485] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.849634] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3d674dfdb6e356885ba0347a8530b2df',container_format='bare',created_at=2024-10-31T11:12:14Z,direct_url=,disk_format='vmdk',id=b9d18b83-05a1-4aa4-87fa-09e4ebd409de,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-379488420-shelved',owner='e74d26cb8f5a47a48dc025b8058c5768',properties=ImageMetaProps,protected=,size=31670784,status='active',tags=,updated_at=2024-10-31T11:12:27Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.849897] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.850083] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.850277] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.850459] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.850618] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.850835] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.850997] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.851225] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.851396] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.851596] env[62183]: DEBUG nova.virt.hardware [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.852460] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbffa99-1360-4856-89eb-3365dd0d3310 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.855835] env[62183]: DEBUG nova.network.neutron [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Successfully created port: 31f6cef9-f14a-4001-a29d-4f02661bbf2d {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.862902] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1906e1f3-301d-48d9-b394-a8afc005ee81 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.878582] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:c5:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '748a5204-8f14-402c-9a6e-f3e6104db082', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37fafdd3-7651-4cfc-818d-2ea8e63ffcb1', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.886054] env[62183]: DEBUG oslo.service.loopingcall [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.886722] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 941.886955] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5a0b64b-d88d-4eb8-a9ef-657eac658600 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.901115] env[62183]: DEBUG nova.compute.manager [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.901977] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0efe11f9-549c-4dfd-b7f7-810339ab2bc0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.911620] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.911620] env[62183]: value = "task-1387261" [ 941.911620] env[62183]: _type = "Task" [ 941.911620] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.921315] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387261, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.932252] env[62183]: DEBUG nova.compute.manager [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 942.230213] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-28d279bd-206e-4ba0-bf84-2980ab41f38d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.230213] env[62183]: DEBUG nova.objects.instance [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lazy-loading 'migration_context' on Instance uuid 28d279bd-206e-4ba0-bf84-2980ab41f38d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.249797] env[62183]: DEBUG nova.scheduler.client.report [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 942.379729] env[62183]: DEBUG nova.network.neutron [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Port b1af05d2-9221-4ac1-a4b4-82f7013cde5c binding to destination host cpu-1 is already ACTIVE {{(pid=62183) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 942.416019] env[62183]: INFO nova.compute.manager [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] instance snapshotting [ 942.416019] env[62183]: DEBUG nova.objects.instance [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'flavor' on Instance uuid a948464b-63aa-4bc8-9885-228049e96d37 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.429026] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387261, 'name': CreateVM_Task, 'duration_secs': 0.364162} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.429026] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 942.429026] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.429026] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.429026] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.429026] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-987331d1-7c3a-480d-88e3-4459eb0f965f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.433326] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 942.433326] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52d2235c-8579-02cb-d48f-d9559a8969ad" [ 942.433326] env[62183]: _type = "Task" [ 942.433326] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.445120] env[62183]: DEBUG nova.compute.manager [req-bb6a0ae1-b7a6-46e6-9810-cfb41600eef3 req-f69ba589-a848-4406-a16e-af14469c6bde service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received event network-changed-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.446238] env[62183]: DEBUG nova.compute.manager [req-bb6a0ae1-b7a6-46e6-9810-cfb41600eef3 req-f69ba589-a848-4406-a16e-af14469c6bde service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Refreshing instance network info cache due to event network-changed-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.446238] env[62183]: DEBUG oslo_concurrency.lockutils [req-bb6a0ae1-b7a6-46e6-9810-cfb41600eef3 req-f69ba589-a848-4406-a16e-af14469c6bde service nova] Acquiring lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.446369] env[62183]: DEBUG oslo_concurrency.lockutils [req-bb6a0ae1-b7a6-46e6-9810-cfb41600eef3 req-f69ba589-a848-4406-a16e-af14469c6bde service nova] Acquired lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.446673] env[62183]: DEBUG nova.network.neutron [req-bb6a0ae1-b7a6-46e6-9810-cfb41600eef3 req-f69ba589-a848-4406-a16e-af14469c6bde service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Refreshing network info cache for port 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 942.451872] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.452264] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Processing image b9d18b83-05a1-4aa4-87fa-09e4ebd409de {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.453252] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de/b9d18b83-05a1-4aa4-87fa-09e4ebd409de.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.453252] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de/b9d18b83-05a1-4aa4-87fa-09e4ebd409de.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.453252] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.454057] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5bd2ed8-dd50-4f96-a996-8a830233d963 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.465019] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.465019] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 942.465019] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4d169be-cedf-4439-993c-e58286749cd0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.469215] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 942.469215] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5299d8b3-2248-7c84-0087-8e83a169b0f4" [ 942.469215] env[62183]: _type = "Task" [ 942.469215] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.478464] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5299d8b3-2248-7c84-0087-8e83a169b0f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.736056] env[62183]: DEBUG nova.objects.base [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Object Instance<28d279bd-206e-4ba0-bf84-2980ab41f38d> lazy-loaded attributes: info_cache,migration_context {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 942.737021] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd0c989-090c-4454-a0ac-eddf5a92c924 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.759702] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.836s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.762474] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.138s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.764359] env[62183]: INFO nova.compute.claims [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.766729] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8eba8c45-759e-482d-9c2b-398100c3acfd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.772466] env[62183]: DEBUG oslo_vmware.api [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 942.772466] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a3105a-a4a3-f2a4-ded9-5a026b33fbd9" [ 942.772466] env[62183]: _type = "Task" [ 942.772466] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.777530] env[62183]: INFO nova.scheduler.client.report [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted allocations for instance 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631 [ 942.781246] env[62183]: DEBUG oslo_vmware.api [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a3105a-a4a3-f2a4-ded9-5a026b33fbd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.923627] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008940d9-9e68-4ef0-a8c2-807c0abbf7b2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.942669] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdb6f9a-8bb8-4986-83d8-eb5db2122d6f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.946156] env[62183]: DEBUG nova.compute.manager [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 942.948132] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 942.950423] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 942.950423] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 942.975311] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.976008] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.976008] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.976008] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.976200] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.976376] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.976550] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.976728] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.976902] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.977081] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.977259] env[62183]: DEBUG nova.virt.hardware [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.978276] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfdb1cd-7985-4def-8734-c36468b4c6ae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.990320] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Preparing fetch location {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 942.990573] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Fetch image to [datastore1] OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6/OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6.vmdk {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 942.990758] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Downloading stream optimized image b9d18b83-05a1-4aa4-87fa-09e4ebd409de to [datastore1] OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6/OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6.vmdk on the data store datastore1 as vApp {{(pid=62183) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 942.990928] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Downloading image file data b9d18b83-05a1-4aa4-87fa-09e4ebd409de to the ESX as VM named 'OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6' {{(pid=62183) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 942.994248] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb2b628-dcfe-45ec-804e-156ce892e5f3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.070974] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 943.070974] env[62183]: value = "resgroup-9" [ 943.070974] env[62183]: _type = "ResourcePool" [ 943.070974] env[62183]: }. {{(pid=62183) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 943.074126] env[62183]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-2421653a-ecd2-446c-beb5-67cd8e26ad1b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.097553] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lease: (returnval){ [ 943.097553] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5224c49a-8d82-937f-95df-f43c90a74bf4" [ 943.097553] env[62183]: _type = "HttpNfcLease" [ 943.097553] env[62183]: } obtained for vApp import into resource pool (val){ [ 943.097553] env[62183]: value = "resgroup-9" [ 943.097553] env[62183]: _type = "ResourcePool" [ 943.097553] env[62183]: }. {{(pid=62183) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 943.098165] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the lease: (returnval){ [ 943.098165] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5224c49a-8d82-937f-95df-f43c90a74bf4" [ 943.098165] env[62183]: _type = "HttpNfcLease" [ 943.098165] env[62183]: } to be ready. {{(pid=62183) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 943.105739] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 943.105739] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5224c49a-8d82-937f-95df-f43c90a74bf4" [ 943.105739] env[62183]: _type = "HttpNfcLease" [ 943.105739] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 943.227174] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "8c4b68d6-fedd-408f-a449-aace7400014a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.227174] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.268331] env[62183]: DEBUG nova.network.neutron [req-bb6a0ae1-b7a6-46e6-9810-cfb41600eef3 req-f69ba589-a848-4406-a16e-af14469c6bde service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updated VIF entry in instance network info cache for port 37fafdd3-7651-4cfc-818d-2ea8e63ffcb1. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 943.268896] env[62183]: DEBUG nova.network.neutron [req-bb6a0ae1-b7a6-46e6-9810-cfb41600eef3 req-f69ba589-a848-4406-a16e-af14469c6bde service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updating instance_info_cache with network_info: [{"id": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "address": "fa:16:3e:3f:c5:6f", "network": {"id": "7d52241a-8c3c-46bc-b64d-1cbc74ec2434", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-104877467-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e74d26cb8f5a47a48dc025b8058c5768", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "748a5204-8f14-402c-9a6e-f3e6104db082", "external-id": "nsx-vlan-transportzone-750", "segmentation_id": 750, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37fafdd3-76", "ovs_interfaceid": "37fafdd3-7651-4cfc-818d-2ea8e63ffcb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.283776] env[62183]: DEBUG oslo_vmware.api [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a3105a-a4a3-f2a4-ded9-5a026b33fbd9, 'name': SearchDatastore_Task, 'duration_secs': 0.014637} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.286425] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.286931] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6a22ebf3-208d-43d3-980a-0ad1e8bf0373 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "7b8a64ad-afcb-4e7c-8c00-86f69d8d3631" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.340s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.402630] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.402973] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.403079] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.459649] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Creating Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 943.459971] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-71765b1a-f5c8-4c73-8182-dbe05686ff67 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.467706] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 943.467706] env[62183]: value = "task-1387263" [ 943.467706] env[62183]: _type = "Task" [ 943.467706] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.475018] env[62183]: DEBUG nova.network.neutron [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Successfully updated port: 31f6cef9-f14a-4001-a29d-4f02661bbf2d {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.482910] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387263, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.606104] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 943.606104] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5224c49a-8d82-937f-95df-f43c90a74bf4" [ 943.606104] env[62183]: _type = "HttpNfcLease" [ 943.606104] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 943.728867] env[62183]: DEBUG nova.compute.manager [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 943.773725] env[62183]: DEBUG oslo_concurrency.lockutils [req-bb6a0ae1-b7a6-46e6-9810-cfb41600eef3 req-f69ba589-a848-4406-a16e-af14469c6bde service nova] Releasing lock "refresh_cache-fb348784-62a0-4d1f-ac7f-f176f3da0dd9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.947492] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fb5f10-6997-45a8-8929-a52cceeafab9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.957791] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80bc4cba-2de1-46dd-b30b-9006054b09e3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.992186] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "refresh_cache-c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.992341] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "refresh_cache-c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.992495] env[62183]: DEBUG nova.network.neutron [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.999578] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52dcd7ba-e448-4966-8eac-055f538b8a0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.008321] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387263, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.011495] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d37336-b1c9-4256-8917-e7c94fe21ff3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.025675] env[62183]: DEBUG nova.compute.provider_tree [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.106940] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 944.106940] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5224c49a-8d82-937f-95df-f43c90a74bf4" [ 944.106940] env[62183]: _type = "HttpNfcLease" [ 944.106940] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 944.248714] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.438128] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.438400] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.438620] env[62183]: DEBUG nova.network.neutron [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 944.471226] env[62183]: DEBUG nova.compute.manager [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Received event network-vif-plugged-31f6cef9-f14a-4001-a29d-4f02661bbf2d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.471459] env[62183]: DEBUG oslo_concurrency.lockutils [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] Acquiring lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.471685] env[62183]: DEBUG oslo_concurrency.lockutils [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] Lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.472914] env[62183]: DEBUG oslo_concurrency.lockutils [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] Lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.472914] env[62183]: DEBUG nova.compute.manager [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] No waiting events found dispatching network-vif-plugged-31f6cef9-f14a-4001-a29d-4f02661bbf2d {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 944.472914] env[62183]: WARNING nova.compute.manager [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Received unexpected event network-vif-plugged-31f6cef9-f14a-4001-a29d-4f02661bbf2d for instance with vm_state building and task_state spawning. [ 944.472914] env[62183]: DEBUG nova.compute.manager [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Received event network-changed-31f6cef9-f14a-4001-a29d-4f02661bbf2d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.472914] env[62183]: DEBUG nova.compute.manager [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Refreshing instance network info cache due to event network-changed-31f6cef9-f14a-4001-a29d-4f02661bbf2d. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 944.472914] env[62183]: DEBUG oslo_concurrency.lockutils [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] Acquiring lock "refresh_cache-c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.491597] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.491744] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquired lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.491892] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Forcefully refreshing network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 944.505847] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387263, 'name': CreateSnapshot_Task, 'duration_secs': 0.869876} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.506132] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Created Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 944.506909] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8862eed-c370-4277-b585-97acf9e2a81b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.529321] env[62183]: DEBUG nova.scheduler.client.report [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 944.536572] env[62183]: DEBUG nova.network.neutron [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 944.608411] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 944.608411] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5224c49a-8d82-937f-95df-f43c90a74bf4" [ 944.608411] env[62183]: _type = "HttpNfcLease" [ 944.608411] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 944.679180] env[62183]: DEBUG nova.network.neutron [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Updating instance_info_cache with network_info: [{"id": "31f6cef9-f14a-4001-a29d-4f02661bbf2d", "address": "fa:16:3e:36:48:e4", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31f6cef9-f1", "ovs_interfaceid": "31f6cef9-f14a-4001-a29d-4f02661bbf2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.776162] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "3a3ca703-2883-4aa9-a33f-326cc25d4838" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.776463] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "3a3ca703-2883-4aa9-a33f-326cc25d4838" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.024203] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Creating linked-clone VM from snapshot {{(pid=62183) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 945.024558] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f6f1aa55-e84e-4f7e-8bab-2c95b6f10a8d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.032889] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 945.032889] env[62183]: value = "task-1387264" [ 945.032889] env[62183]: _type = "Task" [ 945.032889] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.033571] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.034121] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 945.037871] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.390s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.038101] env[62183]: INFO nova.compute.claims [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.054313] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387264, 'name': CloneVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.115286] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 945.115286] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5224c49a-8d82-937f-95df-f43c90a74bf4" [ 945.115286] env[62183]: _type = "HttpNfcLease" [ 945.115286] env[62183]: } is ready. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 945.115286] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 945.115286] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5224c49a-8d82-937f-95df-f43c90a74bf4" [ 945.115286] env[62183]: _type = "HttpNfcLease" [ 945.115286] env[62183]: }. {{(pid=62183) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 945.115286] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f27a9a7-3183-4625-9cde-50e1c8bd9648 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.123310] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520f6910-f670-2f01-9e89-db2c63b10cc0/disk-0.vmdk from lease info. {{(pid=62183) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 945.123883] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Creating HTTP connection to write to file with size = 31670784 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520f6910-f670-2f01-9e89-db2c63b10cc0/disk-0.vmdk. {{(pid=62183) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 945.183859] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "refresh_cache-c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.184183] env[62183]: DEBUG nova.compute.manager [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Instance network_info: |[{"id": "31f6cef9-f14a-4001-a29d-4f02661bbf2d", "address": "fa:16:3e:36:48:e4", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31f6cef9-f1", "ovs_interfaceid": "31f6cef9-f14a-4001-a29d-4f02661bbf2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 945.185642] env[62183]: DEBUG oslo_concurrency.lockutils [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] Acquired lock "refresh_cache-c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.185847] env[62183]: DEBUG nova.network.neutron [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Refreshing network info cache for port 31f6cef9-f14a-4001-a29d-4f02661bbf2d {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 945.187043] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:48:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '205fb402-8eaf-4b61-8f57-8f216024179a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31f6cef9-f14a-4001-a29d-4f02661bbf2d', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.194268] env[62183]: DEBUG oslo.service.loopingcall [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.200508] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.202315] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbb2aa41-1d5a-4f7c-8410-21b36184a072 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.218153] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1bc71ae0-dced-4eb5-9e6d-4f368119fd5e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.226542] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.226542] env[62183]: value = "task-1387265" [ 945.226542] env[62183]: _type = "Task" [ 945.226542] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.236685] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387265, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.237758] env[62183]: DEBUG nova.network.neutron [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance_info_cache with network_info: [{"id": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "address": "fa:16:3e:a4:03:ac", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1af05d2-92", "ovs_interfaceid": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.278360] env[62183]: DEBUG nova.compute.manager [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 945.490260] env[62183]: DEBUG nova.network.neutron [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Updated VIF entry in instance network info cache for port 31f6cef9-f14a-4001-a29d-4f02661bbf2d. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.490714] env[62183]: DEBUG nova.network.neutron [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Updating instance_info_cache with network_info: [{"id": "31f6cef9-f14a-4001-a29d-4f02661bbf2d", "address": "fa:16:3e:36:48:e4", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31f6cef9-f1", "ovs_interfaceid": "31f6cef9-f14a-4001-a29d-4f02661bbf2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.550923] env[62183]: DEBUG nova.compute.utils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.561644] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 945.561932] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 945.564173] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387264, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.610814] env[62183]: DEBUG nova.policy [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b92c15df0ab4e7f9e845e97d3e5912c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '499aba16038147e3b20b06876234cb13', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 945.742045] env[62183]: DEBUG oslo_concurrency.lockutils [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.745749] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387265, 'name': CreateVM_Task, 'duration_secs': 0.343699} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.747900] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.748762] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.748933] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.749271] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 945.749817] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71ea9c1d-0ba4-4f6e-a769-760cbd17100f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.754812] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 945.754812] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52647317-47b6-ee58-bbc5-560510edd024" [ 945.754812] env[62183]: _type = "Task" [ 945.754812] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.763917] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52647317-47b6-ee58-bbc5-560510edd024, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.791537] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.807902] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.925886] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Successfully created port: cea51b38-bd41-4f63-8713-bcc06c8cf941 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 945.994408] env[62183]: DEBUG oslo_concurrency.lockutils [req-5f158caf-6be3-42b4-8291-9888f2788ae1 req-8e6cacd7-b4c8-463d-a6b3-c152382fa7be service nova] Releasing lock "refresh_cache-c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.048314] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387264, 'name': CloneVM_Task} progress is 95%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.053960] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 946.270738] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52647317-47b6-ee58-bbc5-560510edd024, 'name': SearchDatastore_Task, 'duration_secs': 0.026996} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.274400] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.274905] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.275290] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.275627] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.275987] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.277090] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbdb793-35e4-4d48-a3ac-dfd1b89258a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.289487] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-072c1dea-6982-4f84-9260-d8315e7051b0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.309256] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Releasing lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.309478] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updated the network info_cache for instance {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 946.314508] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.315329] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a20139-3a51-46ec-b7f6-f24ff3333ff0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.318626] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6777b937-6687-4c45-9716-0403a0d469cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.321108] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.321534] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.323738] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.323960] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b967fe33-29f9-4ab4-9ab2-aa88703aa057 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.326228] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.327293] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.327530] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.327708] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.327846] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 946.328031] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.333169] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance 'b89455d9-6d00-41ac-95ec-fb036b8c26c9' progress to 83 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 946.344128] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884d6e3e-aaf7-47a7-9e98-d15dc751ec4d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.344734] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 946.344734] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]527e273c-1f1b-1730-b76e-e34627456b13" [ 946.344734] env[62183]: _type = "Task" [ 946.344734] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.380090] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Completed reading data from the image iterator. {{(pid=62183) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 946.381843] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520f6910-f670-2f01-9e89-db2c63b10cc0/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 946.381843] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa51681-a063-4aa9-b8bb-12062e263f0d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.386127] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee8ecc4-5dfc-4c73-abc4-df48aa4674f9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.391839] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527e273c-1f1b-1730-b76e-e34627456b13, 'name': SearchDatastore_Task, 'duration_secs': 0.010939} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.393466] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a90ab6c-983f-4bcb-a897-7c431f39c26f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.399214] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520f6910-f670-2f01-9e89-db2c63b10cc0/disk-0.vmdk is in state: ready. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 946.399391] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520f6910-f670-2f01-9e89-db2c63b10cc0/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 946.400929] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a222b7db-eef7-480f-89da-c170c71d0f75 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.405423] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-3900cbb0-5068-40d9-9872-c11861ba87d3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.408378] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 946.408378] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f31aa9-e436-bfe9-efe3-7a2024f347d2" [ 946.408378] env[62183]: _type = "Task" [ 946.408378] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.421319] env[62183]: DEBUG nova.compute.provider_tree [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.427452] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f31aa9-e436-bfe9-efe3-7a2024f347d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.548772] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387264, 'name': CloneVM_Task, 'duration_secs': 1.374} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.550119] env[62183]: INFO nova.virt.vmwareapi.vmops [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Created linked-clone VM from snapshot [ 946.550246] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b68e27b-4a78-4f00-9a6d-e3fb9fc4789f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.561746] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Uploading image df86a30e-6e83-440e-8c02-085386d00f0d {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 946.581022] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520f6910-f670-2f01-9e89-db2c63b10cc0/disk-0.vmdk. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 946.581265] env[62183]: INFO nova.virt.vmwareapi.images [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Downloaded image file data b9d18b83-05a1-4aa4-87fa-09e4ebd409de [ 946.582111] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f4efcc-759e-48d6-8294-01950ee944d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.601852] env[62183]: DEBUG oslo_vmware.rw_handles [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 946.601852] env[62183]: value = "vm-294489" [ 946.601852] env[62183]: _type = "VirtualMachine" [ 946.601852] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 946.602147] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-01389b0f-698f-44f0-a3ca-59d012484435 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.604009] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-460a7e43-c763-4005-a54b-a6859333e0ac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.611871] env[62183]: DEBUG oslo_vmware.rw_handles [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease: (returnval){ [ 946.611871] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e276f1-e595-89a1-1fcf-e1d85ba530ae" [ 946.611871] env[62183]: _type = "HttpNfcLease" [ 946.611871] env[62183]: } obtained for exporting VM: (result){ [ 946.611871] env[62183]: value = "vm-294489" [ 946.611871] env[62183]: _type = "VirtualMachine" [ 946.611871] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 946.612257] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the lease: (returnval){ [ 946.612257] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e276f1-e595-89a1-1fcf-e1d85ba530ae" [ 946.612257] env[62183]: _type = "HttpNfcLease" [ 946.612257] env[62183]: } to be ready. {{(pid=62183) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 946.618368] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 946.618368] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e276f1-e595-89a1-1fcf-e1d85ba530ae" [ 946.618368] env[62183]: _type = "HttpNfcLease" [ 946.618368] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 946.637858] env[62183]: INFO nova.virt.vmwareapi.images [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] The imported VM was unregistered [ 946.640433] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Caching image {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 946.640678] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Creating directory with path [datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.640952] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c597ffa7-5af5-4d7e-8d24-d4c615506482 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.651456] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Created directory with path [datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.651656] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6/OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6.vmdk to [datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de/b9d18b83-05a1-4aa4-87fa-09e4ebd409de.vmdk. {{(pid=62183) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 946.651920] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c47319df-ba60-4e06-9bed-9b81204082d2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.661038] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 946.661038] env[62183]: value = "task-1387268" [ 946.661038] env[62183]: _type = "Task" [ 946.661038] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.669062] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387268, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.841257] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.847347] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.847678] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ddefaf5-0101-499b-9c96-70c645001761 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.855359] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 946.855359] env[62183]: value = "task-1387269" [ 946.855359] env[62183]: _type = "Task" [ 946.855359] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.864904] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387269, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.922729] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f31aa9-e436-bfe9-efe3-7a2024f347d2, 'name': SearchDatastore_Task, 'duration_secs': 0.024979} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.923070] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.923352] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d/c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 946.924380] env[62183]: DEBUG nova.scheduler.client.report [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 946.927766] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6032772b-9f19-4d5c-a42e-a978116f6739 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.936154] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 946.936154] env[62183]: value = "task-1387270" [ 946.936154] env[62183]: _type = "Task" [ 946.936154] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.947824] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.065792] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 947.090140] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.090456] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.090621] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.090819] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.090992] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.091145] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.091332] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.091496] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.091722] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.091904] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.092123] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.093144] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e1e9b7-ba8d-41aa-a2fd-65f9249919e7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.101675] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705b157e-e10d-4008-82e0-c03c86c244c5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.123337] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 947.123337] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e276f1-e595-89a1-1fcf-e1d85ba530ae" [ 947.123337] env[62183]: _type = "HttpNfcLease" [ 947.123337] env[62183]: } is ready. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 947.123750] env[62183]: DEBUG oslo_vmware.rw_handles [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 947.123750] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e276f1-e595-89a1-1fcf-e1d85ba530ae" [ 947.123750] env[62183]: _type = "HttpNfcLease" [ 947.123750] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 947.124620] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af968cf9-9b46-498d-8c7d-5db227db0b0a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.133273] env[62183]: DEBUG oslo_vmware.rw_handles [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5269b0bf-1732-c337-d953-ce734a615aa1/disk-0.vmdk from lease info. {{(pid=62183) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 947.133508] env[62183]: DEBUG oslo_vmware.rw_handles [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5269b0bf-1732-c337-d953-ce734a615aa1/disk-0.vmdk for reading. {{(pid=62183) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 947.204653] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387268, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.247368] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b1d77439-47cc-41ed-b0cd-ee15f4fc2a8e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.365303] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387269, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.431704] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.432328] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.435103] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.149s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.451509] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.536030] env[62183]: DEBUG nova.compute.manager [req-e18c24f7-ebb2-4ff1-b76e-4d29e196350b req-c53c6600-be38-4f8f-9fdd-949379042594 service nova] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Received event network-vif-plugged-cea51b38-bd41-4f63-8713-bcc06c8cf941 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.536030] env[62183]: DEBUG oslo_concurrency.lockutils [req-e18c24f7-ebb2-4ff1-b76e-4d29e196350b req-c53c6600-be38-4f8f-9fdd-949379042594 service nova] Acquiring lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.536030] env[62183]: DEBUG oslo_concurrency.lockutils [req-e18c24f7-ebb2-4ff1-b76e-4d29e196350b req-c53c6600-be38-4f8f-9fdd-949379042594 service nova] Lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.536030] env[62183]: DEBUG oslo_concurrency.lockutils [req-e18c24f7-ebb2-4ff1-b76e-4d29e196350b req-c53c6600-be38-4f8f-9fdd-949379042594 service nova] Lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.536030] env[62183]: DEBUG nova.compute.manager [req-e18c24f7-ebb2-4ff1-b76e-4d29e196350b req-c53c6600-be38-4f8f-9fdd-949379042594 service nova] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] No waiting events found dispatching network-vif-plugged-cea51b38-bd41-4f63-8713-bcc06c8cf941 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.536030] env[62183]: WARNING nova.compute.manager [req-e18c24f7-ebb2-4ff1-b76e-4d29e196350b req-c53c6600-be38-4f8f-9fdd-949379042594 service nova] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Received unexpected event network-vif-plugged-cea51b38-bd41-4f63-8713-bcc06c8cf941 for instance with vm_state building and task_state spawning. [ 947.646254] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Successfully updated port: cea51b38-bd41-4f63-8713-bcc06c8cf941 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.705152] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387268, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.868278] env[62183]: DEBUG oslo_vmware.api [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387269, 'name': PowerOnVM_Task, 'duration_secs': 0.758402} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.868800] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.869125] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-57b076c8-0e42-4903-946b-19457d2578b8 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance 'b89455d9-6d00-41ac-95ec-fb036b8c26c9' progress to 100 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.942798] env[62183]: DEBUG nova.compute.utils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 947.947661] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 947.947888] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 947.959671] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.003185] env[62183]: DEBUG nova.policy [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b92c15df0ab4e7f9e845e97d3e5912c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '499aba16038147e3b20b06876234cb13', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.148220] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "refresh_cache-ecdd6f2a-80bd-495d-b54a-cf09ace67a87" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.148539] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "refresh_cache-ecdd6f2a-80bd-495d-b54a-cf09ace67a87" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.148810] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.158677] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915e6ce6-96ed-4fbf-99c5-3e15e83bd631 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.169008] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74879647-15fa-4336-986c-784fe70738b6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.210670] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33d4f64-1502-48cf-b8da-fc79ef022c77 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.218810] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387268, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.222565] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3527509-ad95-45ee-a5d0-43c0a5f785d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.237492] env[62183]: DEBUG nova.compute.provider_tree [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.312157] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Successfully created port: 9564d10c-89e8-4c70-b77d-b0c68416935d {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.449144] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.452254] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387270, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.684092] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.716720] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387268, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.741508] env[62183]: DEBUG nova.scheduler.client.report [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 948.834587] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Updating instance_info_cache with network_info: [{"id": "cea51b38-bd41-4f63-8713-bcc06c8cf941", "address": "fa:16:3e:bf:35:c4", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcea51b38-bd", "ovs_interfaceid": "cea51b38-bd41-4f63-8713-bcc06c8cf941", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.949171] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387270, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.224244] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387268, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.337310] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "refresh_cache-ecdd6f2a-80bd-495d-b54a-cf09ace67a87" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.337637] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Instance network_info: |[{"id": "cea51b38-bd41-4f63-8713-bcc06c8cf941", "address": "fa:16:3e:bf:35:c4", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcea51b38-bd", "ovs_interfaceid": "cea51b38-bd41-4f63-8713-bcc06c8cf941", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 949.338087] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:35:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4df917f7-847a-4c0e-b0e3-69a52e4a1554', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cea51b38-bd41-4f63-8713-bcc06c8cf941', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.345952] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Creating folder: Project (499aba16038147e3b20b06876234cb13). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.346238] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b9e6dd52-6161-48ff-bef4-3e86d7e7970f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.357101] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Created folder: Project (499aba16038147e3b20b06876234cb13) in parent group-v294392. [ 949.357292] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Creating folder: Instances. Parent ref: group-v294491. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 949.357526] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a6cb876-5b47-4540-ae09-7ead785a824c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.367533] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Created folder: Instances in parent group-v294491. [ 949.367757] env[62183]: DEBUG oslo.service.loopingcall [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.367944] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 949.368161] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aee94540-a5f1-4d98-a63c-37803c387d3f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.392392] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.392392] env[62183]: value = "task-1387273" [ 949.392392] env[62183]: _type = "Task" [ 949.392392] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.404438] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387273, 'name': CreateVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.450606] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387270, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.321713} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.450958] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d/c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 949.451267] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.451598] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce98164e-289b-4b46-9c0c-d25d61759f76 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.458499] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 949.458499] env[62183]: value = "task-1387274" [ 949.458499] env[62183]: _type = "Task" [ 949.458499] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.462880] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.468320] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387274, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.496105] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.496456] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.496720] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.496949] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.497144] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.497307] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.497581] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.497786] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.497915] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.498108] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.498295] env[62183]: DEBUG nova.virt.hardware [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.499309] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85897eba-6555-4420-b9b0-e0db66089f08 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.508947] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c2c8a9-2ab9-4860-8175-e55ec57ab5a4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.674814] env[62183]: DEBUG nova.compute.manager [req-77f2d3b7-7095-4731-b56f-2acc0f6f08da req-361ef848-8025-4302-a3f3-20fdb7151953 service nova] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Received event network-changed-cea51b38-bd41-4f63-8713-bcc06c8cf941 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.675065] env[62183]: DEBUG nova.compute.manager [req-77f2d3b7-7095-4731-b56f-2acc0f6f08da req-361ef848-8025-4302-a3f3-20fdb7151953 service nova] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Refreshing instance network info cache due to event network-changed-cea51b38-bd41-4f63-8713-bcc06c8cf941. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.675302] env[62183]: DEBUG oslo_concurrency.lockutils [req-77f2d3b7-7095-4731-b56f-2acc0f6f08da req-361ef848-8025-4302-a3f3-20fdb7151953 service nova] Acquiring lock "refresh_cache-ecdd6f2a-80bd-495d-b54a-cf09ace67a87" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.675486] env[62183]: DEBUG oslo_concurrency.lockutils [req-77f2d3b7-7095-4731-b56f-2acc0f6f08da req-361ef848-8025-4302-a3f3-20fdb7151953 service nova] Acquired lock "refresh_cache-ecdd6f2a-80bd-495d-b54a-cf09ace67a87" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.675712] env[62183]: DEBUG nova.network.neutron [req-77f2d3b7-7095-4731-b56f-2acc0f6f08da req-361ef848-8025-4302-a3f3-20fdb7151953 service nova] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Refreshing network info cache for port cea51b38-bd41-4f63-8713-bcc06c8cf941 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.713317] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.713688] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.714039] env[62183]: DEBUG nova.compute.manager [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Going to confirm migration 3 {{(pid=62183) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 949.721541] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387268, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.62643} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.721775] env[62183]: INFO nova.virt.vmwareapi.ds_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6/OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6.vmdk to [datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de/b9d18b83-05a1-4aa4-87fa-09e4ebd409de.vmdk. [ 949.722078] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Cleaning up location [datastore1] OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 949.722175] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_d30ddf8a-929d-4ea7-84b8-e67352a22aa6 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.722510] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a5b30d2-8c52-4b8c-8702-93f9c4aa5e37 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.728747] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 949.728747] env[62183]: value = "task-1387275" [ 949.728747] env[62183]: _type = "Task" [ 949.728747] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.737149] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387275, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.751704] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.317s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.754555] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.506s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.756180] env[62183]: INFO nova.compute.claims [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.904954] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387273, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.969056] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387274, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072077} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.969424] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.970156] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90745ba-2464-4426-9db6-98405c4d4ed7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.991810] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d/c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.992266] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ae2e634-3496-4aaf-8954-85c1ca7979c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.008629] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Successfully updated port: 9564d10c-89e8-4c70-b77d-b0c68416935d {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.016385] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 950.016385] env[62183]: value = "task-1387276" [ 950.016385] env[62183]: _type = "Task" [ 950.016385] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.025662] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387276, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.238411] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387275, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038444} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.238681] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.238888] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de/b9d18b83-05a1-4aa4-87fa-09e4ebd409de.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.239213] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de/b9d18b83-05a1-4aa4-87fa-09e4ebd409de.vmdk to [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9/fb348784-62a0-4d1f-ac7f-f176f3da0dd9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 950.239442] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8702a642-9b1c-4331-af93-7a9185dbd610 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.245932] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 950.245932] env[62183]: value = "task-1387277" [ 950.245932] env[62183]: _type = "Task" [ 950.245932] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.254266] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.255396] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.255570] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquired lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.255757] env[62183]: DEBUG nova.network.neutron [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.255951] env[62183]: DEBUG nova.objects.instance [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lazy-loading 'info_cache' on Instance uuid b89455d9-6d00-41ac-95ec-fb036b8c26c9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.324322] env[62183]: INFO nova.scheduler.client.report [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted allocation for migration b2bb67a2-95bd-43aa-8d40-b021bec5cd50 [ 950.402999] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387273, 'name': CreateVM_Task, 'duration_secs': 0.550722} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.403227] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 950.404076] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.404257] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.404596] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 950.404855] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8eb10d9f-301d-421e-9c07-f752e45e5212 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.409549] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 950.409549] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52da22f3-1719-2a7a-9e50-c036ba3c6515" [ 950.409549] env[62183]: _type = "Task" [ 950.409549] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.413287] env[62183]: DEBUG nova.network.neutron [req-77f2d3b7-7095-4731-b56f-2acc0f6f08da req-361ef848-8025-4302-a3f3-20fdb7151953 service nova] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Updated VIF entry in instance network info cache for port cea51b38-bd41-4f63-8713-bcc06c8cf941. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.413672] env[62183]: DEBUG nova.network.neutron [req-77f2d3b7-7095-4731-b56f-2acc0f6f08da req-361ef848-8025-4302-a3f3-20fdb7151953 service nova] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Updating instance_info_cache with network_info: [{"id": "cea51b38-bd41-4f63-8713-bcc06c8cf941", "address": "fa:16:3e:bf:35:c4", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcea51b38-bd", "ovs_interfaceid": "cea51b38-bd41-4f63-8713-bcc06c8cf941", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.420167] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52da22f3-1719-2a7a-9e50-c036ba3c6515, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.511951] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "refresh_cache-d30cfb69-1814-49b2-a29d-10cfbcc05722" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.512241] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "refresh_cache-d30cfb69-1814-49b2-a29d-10cfbcc05722" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.512448] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.529109] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387276, 'name': ReconfigVM_Task, 'duration_secs': 0.45286} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.529562] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Reconfigured VM instance instance-00000057 to attach disk [datastore1] c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d/c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.530272] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3cfcf01-cb98-4870-956f-72511251e1fd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.537932] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 950.537932] env[62183]: value = "task-1387278" [ 950.537932] env[62183]: _type = "Task" [ 950.537932] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.546674] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387278, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.756030] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387277, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.831126] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c9f519ca-a67a-4443-92e4-2c8c01f65b1e tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.193s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.918203] env[62183]: DEBUG oslo_concurrency.lockutils [req-77f2d3b7-7095-4731-b56f-2acc0f6f08da req-361ef848-8025-4302-a3f3-20fdb7151953 service nova] Releasing lock "refresh_cache-ecdd6f2a-80bd-495d-b54a-cf09ace67a87" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.925076] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52da22f3-1719-2a7a-9e50-c036ba3c6515, 'name': SearchDatastore_Task, 'duration_secs': 0.011028} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.925469] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.925827] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.926062] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.926266] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.926492] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.927092] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba99adaa-d18b-44f6-a71a-635df90a8a17 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.947602] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.947823] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.948646] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de6fd523-5c9a-4cd0-a72c-129eaf411466 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.953931] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 950.953931] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]528f34b7-f918-ff46-0836-554fbd559a51" [ 950.953931] env[62183]: _type = "Task" [ 950.953931] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.960518] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7526dc-a0c6-40a5-89da-7d2fde9570e2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.967603] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528f34b7-f918-ff46-0836-554fbd559a51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.970291] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adab6c39-f6c1-4811-9ddb-fad6110d4855 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.003612] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56402852-7e2b-46f1-a110-888b5e974a89 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.012241] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3b27b2-ae8c-4c12-bce2-616003f788f4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.028815] env[62183]: DEBUG nova.compute.provider_tree [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 951.048254] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387278, 'name': Rename_Task, 'duration_secs': 0.18653} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.048528] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.048863] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab716766-03b0-4371-ac08-3f8e65c61718 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.055263] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 951.055263] env[62183]: value = "task-1387279" [ 951.055263] env[62183]: _type = "Task" [ 951.055263] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.063897] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387279, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.068831] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.258051] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387277, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.465005] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528f34b7-f918-ff46-0836-554fbd559a51, 'name': SearchDatastore_Task, 'duration_secs': 0.080039} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.467025] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e002355b-df8e-4fea-87ef-d48089ca3160 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.471986] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 951.471986] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]526647aa-8004-588f-f2b0-db6ec57bcc48" [ 951.471986] env[62183]: _type = "Task" [ 951.471986] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.483304] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526647aa-8004-588f-f2b0-db6ec57bcc48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.509556] env[62183]: DEBUG nova.network.neutron [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Updating instance_info_cache with network_info: [{"id": "9564d10c-89e8-4c70-b77d-b0c68416935d", "address": "fa:16:3e:2c:35:1b", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9564d10c-89", "ovs_interfaceid": "9564d10c-89e8-4c70-b77d-b0c68416935d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.526557] env[62183]: DEBUG nova.network.neutron [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance_info_cache with network_info: [{"id": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "address": "fa:16:3e:a4:03:ac", "network": {"id": "11b79aba-76ef-46a6-96ea-acca2219bd96", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-70395533-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "594f9c41496146c68921dad4c8d99ec4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496ac502-bfc4-4324-8332-cac473eb7cc4", "external-id": "nsx-vlan-transportzone-415", "segmentation_id": 415, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1af05d2-92", "ovs_interfaceid": "b1af05d2-9221-4ac1-a4b4-82f7013cde5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.552951] env[62183]: ERROR nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [req-ad135b4e-90af-4c4b-8995-8acccc11ba4e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ad135b4e-90af-4c4b-8995-8acccc11ba4e"}]} [ 951.573763] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387279, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.579131] env[62183]: DEBUG nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 951.597737] env[62183]: DEBUG nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 951.599234] env[62183]: DEBUG nova.compute.provider_tree [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 951.613699] env[62183]: DEBUG nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 951.635851] env[62183]: DEBUG nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 951.760419] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387277, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.806793] env[62183]: DEBUG nova.compute.manager [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Received event network-vif-plugged-9564d10c-89e8-4c70-b77d-b0c68416935d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.807390] env[62183]: DEBUG oslo_concurrency.lockutils [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] Acquiring lock "d30cfb69-1814-49b2-a29d-10cfbcc05722-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.807660] env[62183]: DEBUG oslo_concurrency.lockutils [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] Lock "d30cfb69-1814-49b2-a29d-10cfbcc05722-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.807883] env[62183]: DEBUG oslo_concurrency.lockutils [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] Lock "d30cfb69-1814-49b2-a29d-10cfbcc05722-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.808034] env[62183]: DEBUG nova.compute.manager [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] No waiting events found dispatching network-vif-plugged-9564d10c-89e8-4c70-b77d-b0c68416935d {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.808213] env[62183]: WARNING nova.compute.manager [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Received unexpected event network-vif-plugged-9564d10c-89e8-4c70-b77d-b0c68416935d for instance with vm_state building and task_state spawning. [ 951.808378] env[62183]: DEBUG nova.compute.manager [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Received event network-changed-9564d10c-89e8-4c70-b77d-b0c68416935d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.808568] env[62183]: DEBUG nova.compute.manager [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Refreshing instance network info cache due to event network-changed-9564d10c-89e8-4c70-b77d-b0c68416935d. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.808702] env[62183]: DEBUG oslo_concurrency.lockutils [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] Acquiring lock "refresh_cache-d30cfb69-1814-49b2-a29d-10cfbcc05722" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.838903] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1232f9fd-5221-4c21-bc68-3a8ad2a29168 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.848412] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c204be52-8530-43fb-ac0b-f7d6fca36cd3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.881298] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a172688d-9055-4adc-83ee-37957fbcb50c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.889324] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ff9695-fdf6-4a6c-a5f8-a1e69ca0d4c4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.906194] env[62183]: DEBUG nova.compute.provider_tree [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 951.982387] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526647aa-8004-588f-f2b0-db6ec57bcc48, 'name': SearchDatastore_Task, 'duration_secs': 0.085199} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.982800] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.982942] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] ecdd6f2a-80bd-495d-b54a-cf09ace67a87/ecdd6f2a-80bd-495d-b54a-cf09ace67a87.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.983235] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc3335ad-4fc0-4ba3-99c7-d59f658922b7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.990290] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 951.990290] env[62183]: value = "task-1387280" [ 951.990290] env[62183]: _type = "Task" [ 951.990290] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.998512] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387280, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.012231] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "refresh_cache-d30cfb69-1814-49b2-a29d-10cfbcc05722" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.012595] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Instance network_info: |[{"id": "9564d10c-89e8-4c70-b77d-b0c68416935d", "address": "fa:16:3e:2c:35:1b", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9564d10c-89", "ovs_interfaceid": "9564d10c-89e8-4c70-b77d-b0c68416935d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 952.013030] env[62183]: DEBUG oslo_concurrency.lockutils [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] Acquired lock "refresh_cache-d30cfb69-1814-49b2-a29d-10cfbcc05722" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.013229] env[62183]: DEBUG nova.network.neutron [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Refreshing network info cache for port 9564d10c-89e8-4c70-b77d-b0c68416935d {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.014632] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:35:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4df917f7-847a-4c0e-b0e3-69a52e4a1554', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9564d10c-89e8-4c70-b77d-b0c68416935d', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.022562] env[62183]: DEBUG oslo.service.loopingcall [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.026035] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.026928] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c70e7e0-2db0-43d1-a71a-0e3747153f81 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.041883] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Releasing lock "refresh_cache-b89455d9-6d00-41ac-95ec-fb036b8c26c9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.042193] env[62183]: DEBUG nova.objects.instance [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lazy-loading 'migration_context' on Instance uuid b89455d9-6d00-41ac-95ec-fb036b8c26c9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.050052] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.050052] env[62183]: value = "task-1387281" [ 952.050052] env[62183]: _type = "Task" [ 952.050052] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.058266] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387281, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.068314] env[62183]: DEBUG oslo_vmware.api [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387279, 'name': PowerOnVM_Task, 'duration_secs': 0.642817} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.068627] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.068832] env[62183]: INFO nova.compute.manager [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Took 9.12 seconds to spawn the instance on the hypervisor. [ 952.069069] env[62183]: DEBUG nova.compute.manager [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.072010] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc60ebdb-063d-430c-bff5-7ec2c81a957f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.262605] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387277, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.281758] env[62183]: DEBUG oslo_concurrency.lockutils [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "28d279bd-206e-4ba0-bf84-2980ab41f38d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.281862] env[62183]: DEBUG oslo_concurrency.lockutils [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.282084] env[62183]: DEBUG oslo_concurrency.lockutils [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "28d279bd-206e-4ba0-bf84-2980ab41f38d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.282280] env[62183]: DEBUG oslo_concurrency.lockutils [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.282487] env[62183]: DEBUG oslo_concurrency.lockutils [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.285888] env[62183]: INFO nova.compute.manager [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Terminating instance [ 952.288733] env[62183]: DEBUG nova.compute.manager [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.288733] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.289660] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87572be7-d2c7-4ddb-85f3-ebec8254a38f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.295841] env[62183]: DEBUG nova.network.neutron [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Updated VIF entry in instance network info cache for port 9564d10c-89e8-4c70-b77d-b0c68416935d. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.296313] env[62183]: DEBUG nova.network.neutron [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Updating instance_info_cache with network_info: [{"id": "9564d10c-89e8-4c70-b77d-b0c68416935d", "address": "fa:16:3e:2c:35:1b", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9564d10c-89", "ovs_interfaceid": "9564d10c-89e8-4c70-b77d-b0c68416935d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.305054] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.305054] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a88d447-5d19-435e-8aba-06d16c2e1827 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.309989] env[62183]: DEBUG oslo_vmware.api [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 952.309989] env[62183]: value = "task-1387282" [ 952.309989] env[62183]: _type = "Task" [ 952.309989] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.319792] env[62183]: DEBUG oslo_vmware.api [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387282, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.431803] env[62183]: ERROR nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [req-c8b1a7f8-65ee-41dc-b02e-22f9184ff275] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c8b1a7f8-65ee-41dc-b02e-22f9184ff275"}]} [ 952.450035] env[62183]: DEBUG nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 952.465062] env[62183]: DEBUG nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 952.465522] env[62183]: DEBUG nova.compute.provider_tree [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 952.479690] env[62183]: DEBUG nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 952.500804] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387280, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.501939] env[62183]: DEBUG nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 952.545510] env[62183]: DEBUG nova.objects.base [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 952.546892] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da1710e-9f6c-4e0c-866f-1ae4fc928a4f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.586051] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7adcec72-2e41-4e86-aa6c-1c81a988d074 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.593642] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387281, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.598283] env[62183]: INFO nova.compute.manager [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Took 19.22 seconds to build instance. [ 952.601194] env[62183]: DEBUG oslo_vmware.api [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 952.601194] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52973c94-8020-f7a8-bbe3-3d18100f8e3c" [ 952.601194] env[62183]: _type = "Task" [ 952.601194] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.612194] env[62183]: DEBUG oslo_vmware.api [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52973c94-8020-f7a8-bbe3-3d18100f8e3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.749332] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070286d7-55a2-4a0e-b59e-fd2ab56ac24a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.762038] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387277, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.427173} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.766367] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b9d18b83-05a1-4aa4-87fa-09e4ebd409de/b9d18b83-05a1-4aa4-87fa-09e4ebd409de.vmdk to [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9/fb348784-62a0-4d1f-ac7f-f176f3da0dd9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.766367] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0506d8ff-df12-41f2-9e19-6b3fe6977d22 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.770091] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce0faaf-7353-4d0a-99e0-fbcb9c65a09e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.797030] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9/fb348784-62a0-4d1f-ac7f-f176f3da0dd9.vmdk or device None with type streamOptimized {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.824106] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c81d610-56bf-47fa-811e-fb9eae1eb828 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.839163] env[62183]: DEBUG oslo_concurrency.lockutils [req-6f38bbbc-3af8-42dc-9dcd-3f0f35ff8210 req-7be9c1e0-ddb5-4a68-9b6e-85690accbaa5 service nova] Releasing lock "refresh_cache-d30cfb69-1814-49b2-a29d-10cfbcc05722" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.844792] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a497d10d-0bd8-40ad-814a-8df7878e83b9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.853120] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.853322] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.858232] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 952.858232] env[62183]: value = "task-1387283" [ 952.858232] env[62183]: _type = "Task" [ 952.858232] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.864425] env[62183]: DEBUG oslo_vmware.api [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387282, 'name': PowerOffVM_Task, 'duration_secs': 0.31178} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.867840] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 952.868046] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 952.869333] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f823b80f-2f6b-4264-8ab3-f0f8dde7f1d3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.874346] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2114794-b991-42d9-a2a2-f579f924d127 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.882804] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387283, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.892444] env[62183]: DEBUG nova.compute.provider_tree [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 952.965173] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 952.965493] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 952.965648] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleting the datastore file [datastore2] 28d279bd-206e-4ba0-bf84-2980ab41f38d {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 952.966016] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e957b851-8e41-43c0-8ac9-445ae04a69ab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.973981] env[62183]: DEBUG oslo_vmware.api [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 952.973981] env[62183]: value = "task-1387285" [ 952.973981] env[62183]: _type = "Task" [ 952.973981] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.984139] env[62183]: DEBUG oslo_vmware.api [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387285, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.003393] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387280, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.062777] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387281, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.102654] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc96fea2-45f7-47ab-b6ba-94e8f7fb264f tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.732s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.111780] env[62183]: DEBUG oslo_vmware.api [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52973c94-8020-f7a8-bbe3-3d18100f8e3c, 'name': SearchDatastore_Task, 'duration_secs': 0.034128} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.112115] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.356982] env[62183]: DEBUG nova.compute.manager [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 953.377818] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387283, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.440512] env[62183]: DEBUG nova.scheduler.client.report [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 112 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 953.440917] env[62183]: DEBUG nova.compute.provider_tree [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 112 to 113 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 953.441176] env[62183]: DEBUG nova.compute.provider_tree [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 953.489741] env[62183]: DEBUG oslo_vmware.api [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387285, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.507447] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387280, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.564323] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387281, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.860355] env[62183]: DEBUG nova.compute.manager [req-22acffa2-0b05-4839-b651-47e3a0c8a23d req-58cd29cc-e230-4c49-8f88-3bfbdf9a891c service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Received event network-changed-31f6cef9-f14a-4001-a29d-4f02661bbf2d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.860837] env[62183]: DEBUG nova.compute.manager [req-22acffa2-0b05-4839-b651-47e3a0c8a23d req-58cd29cc-e230-4c49-8f88-3bfbdf9a891c service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Refreshing instance network info cache due to event network-changed-31f6cef9-f14a-4001-a29d-4f02661bbf2d. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 953.860837] env[62183]: DEBUG oslo_concurrency.lockutils [req-22acffa2-0b05-4839-b651-47e3a0c8a23d req-58cd29cc-e230-4c49-8f88-3bfbdf9a891c service nova] Acquiring lock "refresh_cache-c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.861111] env[62183]: DEBUG oslo_concurrency.lockutils [req-22acffa2-0b05-4839-b651-47e3a0c8a23d req-58cd29cc-e230-4c49-8f88-3bfbdf9a891c service nova] Acquired lock "refresh_cache-c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.861111] env[62183]: DEBUG nova.network.neutron [req-22acffa2-0b05-4839-b651-47e3a0c8a23d req-58cd29cc-e230-4c49-8f88-3bfbdf9a891c service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Refreshing network info cache for port 31f6cef9-f14a-4001-a29d-4f02661bbf2d {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.880929] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387283, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.883171] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.948854] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.194s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.950074] env[62183]: DEBUG nova.compute.manager [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.953918] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.146s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.955624] env[62183]: INFO nova.compute.claims [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 953.985086] env[62183]: DEBUG oslo_vmware.api [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387285, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.763461} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.985559] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.985559] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.985726] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.985963] env[62183]: INFO nova.compute.manager [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Took 1.70 seconds to destroy the instance on the hypervisor. [ 953.986240] env[62183]: DEBUG oslo.service.loopingcall [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.986461] env[62183]: DEBUG nova.compute.manager [-] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.986554] env[62183]: DEBUG nova.network.neutron [-] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 954.002935] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387280, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.572691} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.003916] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] ecdd6f2a-80bd-495d-b54a-cf09ace67a87/ecdd6f2a-80bd-495d-b54a-cf09ace67a87.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 954.003916] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.003916] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6469281d-0d2d-461b-91cf-4f484fe030a3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.010091] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 954.010091] env[62183]: value = "task-1387286" [ 954.010091] env[62183]: _type = "Task" [ 954.010091] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.018177] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.063048] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387281, 'name': CreateVM_Task, 'duration_secs': 1.719794} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.064023] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 954.064023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.064271] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.064662] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 954.064906] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8699ee4b-5368-4405-b1de-f817961290f7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.070324] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 954.070324] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e1f2f9-2b57-d6a0-ef81-2a77a0be3bb5" [ 954.070324] env[62183]: _type = "Task" [ 954.070324] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.078125] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e1f2f9-2b57-d6a0-ef81-2a77a0be3bb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.276879] env[62183]: DEBUG nova.compute.manager [req-f573e1be-12e9-4ec5-8458-d7ea0e5c642b req-a8787613-002b-4312-9360-5a6d210e0001 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Received event network-vif-deleted-f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.277163] env[62183]: INFO nova.compute.manager [req-f573e1be-12e9-4ec5-8458-d7ea0e5c642b req-a8787613-002b-4312-9360-5a6d210e0001 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Neutron deleted interface f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc; detaching it from the instance and deleting it from the info cache [ 954.277350] env[62183]: DEBUG nova.network.neutron [req-f573e1be-12e9-4ec5-8458-d7ea0e5c642b req-a8787613-002b-4312-9360-5a6d210e0001 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.378370] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387283, 'name': ReconfigVM_Task, 'duration_secs': 1.119487} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.378615] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Reconfigured VM instance instance-0000004a to attach disk [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9/fb348784-62a0-4d1f-ac7f-f176f3da0dd9.vmdk or device None with type streamOptimized {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.380282] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fcf3a955-fa80-4f4d-a3c9-fedaa4bfda02 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.386527] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 954.386527] env[62183]: value = "task-1387287" [ 954.386527] env[62183]: _type = "Task" [ 954.386527] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.394413] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387287, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.462395] env[62183]: DEBUG nova.compute.utils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 954.463814] env[62183]: DEBUG nova.compute.manager [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 954.463984] env[62183]: DEBUG nova.network.neutron [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 954.506303] env[62183]: DEBUG nova.policy [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbf2ac3d1b8f46cf89df0e7a106c75e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ffc6a4c3a1e74a9b8d9eae70dce238a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 954.522597] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.585941] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e1f2f9-2b57-d6a0-ef81-2a77a0be3bb5, 'name': SearchDatastore_Task, 'duration_secs': 0.019007} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.588682] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.588952] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.589260] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.589427] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.589637] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.589940] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f863d6e-0d38-4f08-b5e7-0149ec962b4d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.600565] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.600880] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.601773] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7749e4d6-aa86-4dd9-b7cc-ff59a76dd7b9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.607603] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 954.607603] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e4a883-713f-4037-cf13-14f667594d95" [ 954.607603] env[62183]: _type = "Task" [ 954.607603] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.616992] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e4a883-713f-4037-cf13-14f667594d95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.629221] env[62183]: DEBUG nova.network.neutron [req-22acffa2-0b05-4839-b651-47e3a0c8a23d req-58cd29cc-e230-4c49-8f88-3bfbdf9a891c service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Updated VIF entry in instance network info cache for port 31f6cef9-f14a-4001-a29d-4f02661bbf2d. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 954.629579] env[62183]: DEBUG nova.network.neutron [req-22acffa2-0b05-4839-b651-47e3a0c8a23d req-58cd29cc-e230-4c49-8f88-3bfbdf9a891c service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Updating instance_info_cache with network_info: [{"id": "31f6cef9-f14a-4001-a29d-4f02661bbf2d", "address": "fa:16:3e:36:48:e4", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31f6cef9-f1", "ovs_interfaceid": "31f6cef9-f14a-4001-a29d-4f02661bbf2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.746156] env[62183]: DEBUG nova.network.neutron [-] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.780159] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4cd7cb9-1332-4b81-a472-5ac8dd7189c4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.789673] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8370ff6a-85c6-4301-a6b4-3f1c7ec17df1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.802447] env[62183]: DEBUG nova.network.neutron [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Successfully created port: e301232c-2b72-4bc5-b5f5-811b208bcc94 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.823375] env[62183]: DEBUG nova.compute.manager [req-f573e1be-12e9-4ec5-8458-d7ea0e5c642b req-a8787613-002b-4312-9360-5a6d210e0001 service nova] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Detach interface failed, port_id=f2ec12e2-ad9a-45fa-a8d7-ebbd3474f7fc, reason: Instance 28d279bd-206e-4ba0-bf84-2980ab41f38d could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 954.897359] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387287, 'name': Rename_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.969162] env[62183]: DEBUG nova.compute.manager [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 955.025998] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387286, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.002474} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.026346] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.028176] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763951a2-dfd1-44af-9bc9-c271fc956661 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.056852] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] ecdd6f2a-80bd-495d-b54a-cf09ace67a87/ecdd6f2a-80bd-495d-b54a-cf09ace67a87.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.060570] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d94e8ce7-f48a-46e1-984f-6c4925f8559a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.087184] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 955.087184] env[62183]: value = "task-1387288" [ 955.087184] env[62183]: _type = "Task" [ 955.087184] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.099662] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387288, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.119764] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e4a883-713f-4037-cf13-14f667594d95, 'name': SearchDatastore_Task, 'duration_secs': 0.009946} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.124146] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfd887e7-310b-4de9-b183-34174bc5b92f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.128682] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 955.128682] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52d31b41-f930-adcf-e0f3-ebb5038bc165" [ 955.128682] env[62183]: _type = "Task" [ 955.128682] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.132009] env[62183]: DEBUG oslo_concurrency.lockutils [req-22acffa2-0b05-4839-b651-47e3a0c8a23d req-58cd29cc-e230-4c49-8f88-3bfbdf9a891c service nova] Releasing lock "refresh_cache-c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.139647] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d31b41-f930-adcf-e0f3-ebb5038bc165, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.244939] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc66f2d1-5d52-496d-ad51-9804c1322d68 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.248644] env[62183]: INFO nova.compute.manager [-] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Took 1.26 seconds to deallocate network for instance. [ 955.258101] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c6541f-c545-4595-8fb6-d49d55bd0aa7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.293088] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295c570e-a821-4fb2-a9d0-65af86b4704e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.301986] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04a25c6-d8fa-4478-9762-f3c9f19e91d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.320364] env[62183]: DEBUG nova.compute.provider_tree [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.398028] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387287, 'name': Rename_Task, 'duration_secs': 0.686516} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.398360] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 955.398627] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13911a14-9345-44c6-9122-df7355c4ebff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.405491] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 955.405491] env[62183]: value = "task-1387289" [ 955.405491] env[62183]: _type = "Task" [ 955.405491] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.413678] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387289, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.599421] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387288, 'name': ReconfigVM_Task, 'duration_secs': 0.487012} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.599842] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Reconfigured VM instance instance-00000058 to attach disk [datastore1] ecdd6f2a-80bd-495d-b54a-cf09ace67a87/ecdd6f2a-80bd-495d-b54a-cf09ace67a87.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.600697] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65ca3f13-87d8-467a-be5a-1fd75be3f8ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.607768] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 955.607768] env[62183]: value = "task-1387290" [ 955.607768] env[62183]: _type = "Task" [ 955.607768] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.618775] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387290, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.638400] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d31b41-f930-adcf-e0f3-ebb5038bc165, 'name': SearchDatastore_Task, 'duration_secs': 0.009807} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.638724] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.639009] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] d30cfb69-1814-49b2-a29d-10cfbcc05722/d30cfb69-1814-49b2-a29d-10cfbcc05722.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.639323] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1dbf0e63-b3f3-489f-910b-dcb8abcf2b57 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.645857] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 955.645857] env[62183]: value = "task-1387291" [ 955.645857] env[62183]: _type = "Task" [ 955.645857] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.654351] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.759018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.823309] env[62183]: DEBUG nova.scheduler.client.report [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 955.919090] env[62183]: DEBUG oslo_vmware.api [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387289, 'name': PowerOnVM_Task, 'duration_secs': 0.472312} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.919090] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 955.980604] env[62183]: DEBUG nova.compute.manager [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 956.034845] env[62183]: DEBUG nova.compute.manager [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.036107] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b4d9f8-c5c2-445f-ad18-afd455b10472 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.058797] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.059077] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.059318] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.059576] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.059576] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.059755] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.059944] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.060122] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.060301] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.060511] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.060861] env[62183]: DEBUG nova.virt.hardware [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.061878] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f17ea6e-f878-476a-8103-fba01f1d18d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.070674] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c44062c-6eb3-4b66-80fb-0091a1eff87d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.117025] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387290, 'name': Rename_Task, 'duration_secs': 0.17197} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.117346] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.117603] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7eb7bdf8-ed7a-4dd1-8ead-01d29fffbe39 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.124018] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 956.124018] env[62183]: value = "task-1387292" [ 956.124018] env[62183]: _type = "Task" [ 956.124018] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.132760] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387292, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.156145] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387291, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.329588] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.330237] env[62183]: DEBUG nova.compute.manager [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 956.334333] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 9.493s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.334550] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.334776] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 956.335182] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.223s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.337498] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737ba013-334f-40f6-9b27-8917c76dccd3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.347281] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e00b7be-92fd-4f52-b521-f0c5ec85a1f6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.362896] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44d959f-3529-4d19-a721-0d2040d4a634 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.370560] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8ec0f5-c279-480b-8104-0e664739a685 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.403035] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179937MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 956.403201] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.466885] env[62183]: DEBUG nova.compute.manager [req-01963f0c-b5ce-47e7-8800-ac4a95c34494 req-2ae0c33a-6346-4daf-b1bb-5020bcad4059 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Received event network-vif-plugged-e301232c-2b72-4bc5-b5f5-811b208bcc94 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.467091] env[62183]: DEBUG oslo_concurrency.lockutils [req-01963f0c-b5ce-47e7-8800-ac4a95c34494 req-2ae0c33a-6346-4daf-b1bb-5020bcad4059 service nova] Acquiring lock "8c4b68d6-fedd-408f-a449-aace7400014a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.467318] env[62183]: DEBUG oslo_concurrency.lockutils [req-01963f0c-b5ce-47e7-8800-ac4a95c34494 req-2ae0c33a-6346-4daf-b1bb-5020bcad4059 service nova] Lock "8c4b68d6-fedd-408f-a449-aace7400014a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.467489] env[62183]: DEBUG oslo_concurrency.lockutils [req-01963f0c-b5ce-47e7-8800-ac4a95c34494 req-2ae0c33a-6346-4daf-b1bb-5020bcad4059 service nova] Lock "8c4b68d6-fedd-408f-a449-aace7400014a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.467658] env[62183]: DEBUG nova.compute.manager [req-01963f0c-b5ce-47e7-8800-ac4a95c34494 req-2ae0c33a-6346-4daf-b1bb-5020bcad4059 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] No waiting events found dispatching network-vif-plugged-e301232c-2b72-4bc5-b5f5-811b208bcc94 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 956.467869] env[62183]: WARNING nova.compute.manager [req-01963f0c-b5ce-47e7-8800-ac4a95c34494 req-2ae0c33a-6346-4daf-b1bb-5020bcad4059 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Received unexpected event network-vif-plugged-e301232c-2b72-4bc5-b5f5-811b208bcc94 for instance with vm_state building and task_state spawning. [ 956.548334] env[62183]: DEBUG nova.network.neutron [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Successfully updated port: e301232c-2b72-4bc5-b5f5-811b208bcc94 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.561129] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab715185-a38f-47d5-aff5-0dd96ac43a44 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 26.540s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.634148] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387292, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.656514] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387291, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.601803} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.656827] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] d30cfb69-1814-49b2-a29d-10cfbcc05722/d30cfb69-1814-49b2-a29d-10cfbcc05722.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.657141] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.657438] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c081fec-dff7-46dc-916d-0d6ace90ecce {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.664865] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 956.664865] env[62183]: value = "task-1387293" [ 956.664865] env[62183]: _type = "Task" [ 956.664865] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.672909] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.838435] env[62183]: DEBUG nova.compute.utils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.839862] env[62183]: DEBUG nova.compute.manager [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 956.840051] env[62183]: DEBUG nova.network.neutron [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 956.883724] env[62183]: DEBUG nova.policy [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88efcb2057e84413806f9ed809c260e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eda7e297c551407295b3c67fa0f1dbc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.049056] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ba3e66-8728-48f4-a73a-d453ddea5543 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.052396] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.052540] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.052687] env[62183]: DEBUG nova.network.neutron [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.057374] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254aa617-6790-4009-9fdb-1810e644aa27 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.089322] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c522438a-e004-4bdc-81e5-530e9cfd5f5b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.096894] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39095d2-f91c-4e8f-8752-4384c34b74d7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.110289] env[62183]: DEBUG nova.compute.provider_tree [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.133253] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387292, 'name': PowerOnVM_Task, 'duration_secs': 0.899335} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.133564] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.133831] env[62183]: INFO nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Took 10.07 seconds to spawn the instance on the hypervisor. [ 957.133963] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.134765] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a98ef9-0c0d-474b-af41-64ae13a32b2d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.156369] env[62183]: DEBUG nova.network.neutron [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Successfully created port: 21c40087-5244-413a-ae18-e7e27f2abc34 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 957.174744] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069536} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.175057] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.175859] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e34732-ae14-419e-8234-a7b1c7855034 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.198143] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] d30cfb69-1814-49b2-a29d-10cfbcc05722/d30cfb69-1814-49b2-a29d-10cfbcc05722.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.198670] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1882356e-2b6b-4f75-8d09-f316c247ed47 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.217514] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 957.217514] env[62183]: value = "task-1387294" [ 957.217514] env[62183]: _type = "Task" [ 957.217514] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.225551] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387294, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.343082] env[62183]: DEBUG nova.compute.manager [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 957.594065] env[62183]: DEBUG nova.network.neutron [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 957.614557] env[62183]: DEBUG nova.scheduler.client.report [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 957.656349] env[62183]: INFO nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Took 21.05 seconds to build instance. [ 957.728893] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387294, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.784720] env[62183]: DEBUG nova.network.neutron [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updating instance_info_cache with network_info: [{"id": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "address": "fa:16:3e:10:9e:d2", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape301232c-2b", "ovs_interfaceid": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.158893] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.566s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.229094] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387294, 'name': ReconfigVM_Task, 'duration_secs': 0.653959} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.229372] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Reconfigured VM instance instance-00000059 to attach disk [datastore1] d30cfb69-1814-49b2-a29d-10cfbcc05722/d30cfb69-1814-49b2-a29d-10cfbcc05722.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.229992] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5ca0b1da-9d20-4a85-bafd-83a5673695b2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.236448] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 958.236448] env[62183]: value = "task-1387295" [ 958.236448] env[62183]: _type = "Task" [ 958.236448] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.244088] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387295, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.288457] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Releasing lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.288817] env[62183]: DEBUG nova.compute.manager [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Instance network_info: |[{"id": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "address": "fa:16:3e:10:9e:d2", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape301232c-2b", "ovs_interfaceid": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.289303] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:9e:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbdab640-5fea-4254-8bd3-f855b7eaca0d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e301232c-2b72-4bc5-b5f5-811b208bcc94', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.296837] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Creating folder: Project (ffc6a4c3a1e74a9b8d9eae70dce238a0). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 958.297255] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d480e554-6134-4bd1-bf8f-7cd2574143c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.308652] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Created folder: Project (ffc6a4c3a1e74a9b8d9eae70dce238a0) in parent group-v294392. [ 958.308879] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Creating folder: Instances. Parent ref: group-v294495. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 958.309178] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-72bc8e89-e9b0-4c19-a8eb-cdf0422fb49b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.317837] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Created folder: Instances in parent group-v294495. [ 958.318077] env[62183]: DEBUG oslo.service.loopingcall [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.318275] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.318476] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d203d154-ed87-47f5-bb26-d0cca89dab3f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.336287] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.336287] env[62183]: value = "task-1387298" [ 958.336287] env[62183]: _type = "Task" [ 958.336287] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.343450] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387298, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.352710] env[62183]: DEBUG nova.compute.manager [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 958.376043] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.376357] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.376550] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.376822] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.376986] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.377154] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.377365] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.377528] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.377700] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.377877] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.378051] env[62183]: DEBUG nova.virt.hardware [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.378932] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af833c28-e1e5-4912-9b79-9721b62da5cb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.386495] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d1ff50-34ed-40a1-8560-f97601ae580d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.628893] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.292s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.631035] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.747s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.631975] env[62183]: INFO nova.compute.claims [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.641678] env[62183]: DEBUG nova.compute.manager [req-39a79ef6-9c64-44ef-87d0-0973ca701275 req-d206b807-dbad-417c-a950-54ee98a39276 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Received event network-changed-e301232c-2b72-4bc5-b5f5-811b208bcc94 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.641884] env[62183]: DEBUG nova.compute.manager [req-39a79ef6-9c64-44ef-87d0-0973ca701275 req-d206b807-dbad-417c-a950-54ee98a39276 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Refreshing instance network info cache due to event network-changed-e301232c-2b72-4bc5-b5f5-811b208bcc94. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.642127] env[62183]: DEBUG oslo_concurrency.lockutils [req-39a79ef6-9c64-44ef-87d0-0973ca701275 req-d206b807-dbad-417c-a950-54ee98a39276 service nova] Acquiring lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.642280] env[62183]: DEBUG oslo_concurrency.lockutils [req-39a79ef6-9c64-44ef-87d0-0973ca701275 req-d206b807-dbad-417c-a950-54ee98a39276 service nova] Acquired lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.642448] env[62183]: DEBUG nova.network.neutron [req-39a79ef6-9c64-44ef-87d0-0973ca701275 req-d206b807-dbad-417c-a950-54ee98a39276 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Refreshing network info cache for port e301232c-2b72-4bc5-b5f5-811b208bcc94 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 958.747453] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387295, 'name': Rename_Task, 'duration_secs': 0.223739} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.747726] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.747976] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3bcf844-688c-41ca-9fd2-da3a858a111b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.755077] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 958.755077] env[62183]: value = "task-1387299" [ 958.755077] env[62183]: _type = "Task" [ 958.755077] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.762897] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387299, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.814418] env[62183]: DEBUG oslo_vmware.rw_handles [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5269b0bf-1732-c337-d953-ce734a615aa1/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 958.816102] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4536d633-2913-463c-89db-2c6a52ea9c30 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.822834] env[62183]: DEBUG oslo_vmware.rw_handles [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5269b0bf-1732-c337-d953-ce734a615aa1/disk-0.vmdk is in state: ready. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 958.823021] env[62183]: ERROR oslo_vmware.rw_handles [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5269b0bf-1732-c337-d953-ce734a615aa1/disk-0.vmdk due to incomplete transfer. [ 958.823250] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1388fe54-9e43-4c8a-9983-ebc996ed3c5d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.829999] env[62183]: DEBUG oslo_vmware.rw_handles [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5269b0bf-1732-c337-d953-ce734a615aa1/disk-0.vmdk. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 958.830254] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Uploaded image df86a30e-6e83-440e-8c02-085386d00f0d to the Glance image server {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 958.832566] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Destroying the VM {{(pid=62183) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 958.832823] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6422e467-4a11-4cf0-ae1b-886459e7967e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.840689] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 958.840689] env[62183]: value = "task-1387300" [ 958.840689] env[62183]: _type = "Task" [ 958.840689] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.846917] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387298, 'name': CreateVM_Task, 'duration_secs': 0.470236} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.847673] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.848448] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.848622] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.848939] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.851770] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7997db9d-42d7-43f1-998b-85904552c900 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.853331] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387300, 'name': Destroy_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.856193] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 958.856193] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5276c97b-3c42-ce3d-6bc1-0a3a37838db7" [ 958.856193] env[62183]: _type = "Task" [ 958.856193] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.863445] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5276c97b-3c42-ce3d-6bc1-0a3a37838db7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.929089] env[62183]: DEBUG nova.network.neutron [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Successfully updated port: 21c40087-5244-413a-ae18-e7e27f2abc34 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 959.195351] env[62183]: INFO nova.scheduler.client.report [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted allocation for migration bd8cf8f4-2aae-4118-8c5a-2cdbfb615009 [ 959.265236] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387299, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.347363] env[62183]: DEBUG nova.network.neutron [req-39a79ef6-9c64-44ef-87d0-0973ca701275 req-d206b807-dbad-417c-a950-54ee98a39276 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updated VIF entry in instance network info cache for port e301232c-2b72-4bc5-b5f5-811b208bcc94. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 959.347729] env[62183]: DEBUG nova.network.neutron [req-39a79ef6-9c64-44ef-87d0-0973ca701275 req-d206b807-dbad-417c-a950-54ee98a39276 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updating instance_info_cache with network_info: [{"id": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "address": "fa:16:3e:10:9e:d2", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape301232c-2b", "ovs_interfaceid": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.352092] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387300, 'name': Destroy_Task} progress is 33%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.365864] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5276c97b-3c42-ce3d-6bc1-0a3a37838db7, 'name': SearchDatastore_Task, 'duration_secs': 0.017344} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.366170] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.366414] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.366648] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.366795] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.366976] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.367253] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8518373-f747-47c9-ab3f-6c06ffb70fe6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.374804] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.374974] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.375702] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d701d93-b81a-45b8-b005-796463510d02 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.380438] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 959.380438] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52858390-7ba7-e71e-775a-5d7881de95c1" [ 959.380438] env[62183]: _type = "Task" [ 959.380438] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.388164] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52858390-7ba7-e71e-775a-5d7881de95c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.432503] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "refresh_cache-3a3ca703-2883-4aa9-a33f-326cc25d4838" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.432503] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "refresh_cache-3a3ca703-2883-4aa9-a33f-326cc25d4838" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.432503] env[62183]: DEBUG nova.network.neutron [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 959.703502] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.990s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.767837] env[62183]: DEBUG oslo_vmware.api [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387299, 'name': PowerOnVM_Task, 'duration_secs': 0.804983} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.768125] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.768381] env[62183]: INFO nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Took 10.31 seconds to spawn the instance on the hypervisor. [ 959.768574] env[62183]: DEBUG nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 959.769364] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-543627a8-63d5-4ef1-a8a2-59866c36a3c8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.822601] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d84325-5401-4913-aaa5-9ee05251cc32 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.830872] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c72d695-2a5d-4cc0-b961-b5e3a0646919 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.860239] env[62183]: DEBUG oslo_concurrency.lockutils [req-39a79ef6-9c64-44ef-87d0-0973ca701275 req-d206b807-dbad-417c-a950-54ee98a39276 service nova] Releasing lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.864118] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e549fc01-2a70-4a77-a1ef-b6e2c9211fbc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.871573] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387300, 'name': Destroy_Task, 'duration_secs': 0.704239} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.873643] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Destroyed the VM [ 959.873905] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deleting Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 959.874224] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c1cfe980-960a-4db9-8c47-2e00cd4fe95d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.876762] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435f6cd8-2da2-41c3-8041-669992ace956 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.891839] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 959.891839] env[62183]: value = "task-1387301" [ 959.891839] env[62183]: _type = "Task" [ 959.891839] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.892305] env[62183]: DEBUG nova.compute.provider_tree [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.900790] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52858390-7ba7-e71e-775a-5d7881de95c1, 'name': SearchDatastore_Task, 'duration_secs': 0.014719} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.902503] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec58390f-ad90-4d1b-9a77-d58c7799e892 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.907555] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387301, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.911014] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 959.911014] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52d1ac65-5f3d-7ab0-0b70-36da43903051" [ 959.911014] env[62183]: _type = "Task" [ 959.911014] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.920281] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d1ac65-5f3d-7ab0-0b70-36da43903051, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.983167] env[62183]: DEBUG nova.network.neutron [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 960.189418] env[62183]: DEBUG nova.network.neutron [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Updating instance_info_cache with network_info: [{"id": "21c40087-5244-413a-ae18-e7e27f2abc34", "address": "fa:16:3e:92:b7:b9", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21c40087-52", "ovs_interfaceid": "21c40087-5244-413a-ae18-e7e27f2abc34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.290342] env[62183]: INFO nova.compute.manager [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Took 23.66 seconds to build instance. [ 960.395758] env[62183]: DEBUG nova.scheduler.client.report [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 960.407346] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387301, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.420754] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d1ac65-5f3d-7ab0-0b70-36da43903051, 'name': SearchDatastore_Task, 'duration_secs': 0.0122} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.421010] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.421269] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 8c4b68d6-fedd-408f-a449-aace7400014a/8c4b68d6-fedd-408f-a449-aace7400014a.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.421508] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1fa424a7-03e0-4a6c-ab3f-fec33359b4c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.427893] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 960.427893] env[62183]: value = "task-1387302" [ 960.427893] env[62183]: _type = "Task" [ 960.427893] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.435159] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387302, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.677599] env[62183]: DEBUG nova.compute.manager [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Received event network-vif-plugged-21c40087-5244-413a-ae18-e7e27f2abc34 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.677858] env[62183]: DEBUG oslo_concurrency.lockutils [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] Acquiring lock "3a3ca703-2883-4aa9-a33f-326cc25d4838-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.678054] env[62183]: DEBUG oslo_concurrency.lockutils [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] Lock "3a3ca703-2883-4aa9-a33f-326cc25d4838-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.678248] env[62183]: DEBUG oslo_concurrency.lockutils [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] Lock "3a3ca703-2883-4aa9-a33f-326cc25d4838-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.678426] env[62183]: DEBUG nova.compute.manager [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] No waiting events found dispatching network-vif-plugged-21c40087-5244-413a-ae18-e7e27f2abc34 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.678801] env[62183]: WARNING nova.compute.manager [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Received unexpected event network-vif-plugged-21c40087-5244-413a-ae18-e7e27f2abc34 for instance with vm_state building and task_state spawning. [ 960.678801] env[62183]: DEBUG nova.compute.manager [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Received event network-changed-21c40087-5244-413a-ae18-e7e27f2abc34 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.679024] env[62183]: DEBUG nova.compute.manager [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Refreshing instance network info cache due to event network-changed-21c40087-5244-413a-ae18-e7e27f2abc34. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 960.679288] env[62183]: DEBUG oslo_concurrency.lockutils [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] Acquiring lock "refresh_cache-3a3ca703-2883-4aa9-a33f-326cc25d4838" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.692128] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "refresh_cache-3a3ca703-2883-4aa9-a33f-326cc25d4838" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.692405] env[62183]: DEBUG nova.compute.manager [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Instance network_info: |[{"id": "21c40087-5244-413a-ae18-e7e27f2abc34", "address": "fa:16:3e:92:b7:b9", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21c40087-52", "ovs_interfaceid": "21c40087-5244-413a-ae18-e7e27f2abc34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 960.692698] env[62183]: DEBUG oslo_concurrency.lockutils [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] Acquired lock "refresh_cache-3a3ca703-2883-4aa9-a33f-326cc25d4838" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.692957] env[62183]: DEBUG nova.network.neutron [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Refreshing network info cache for port 21c40087-5244-413a-ae18-e7e27f2abc34 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 960.694357] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:b7:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21c40087-5244-413a-ae18-e7e27f2abc34', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 960.701924] env[62183]: DEBUG oslo.service.loopingcall [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.703422] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 960.703697] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce28236c-e45e-45e7-ae3e-92a83d259e81 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.724484] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 960.724484] env[62183]: value = "task-1387303" [ 960.724484] env[62183]: _type = "Task" [ 960.724484] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.732818] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387303, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.738751] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.739036] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.739269] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.739475] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.739649] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.741901] env[62183]: INFO nova.compute.manager [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Terminating instance [ 960.743858] env[62183]: DEBUG nova.compute.manager [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.744159] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.744913] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c04b76-f6c5-47ba-9634-dd4d12ee9af4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.753995] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.754261] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ae09aec-10bc-46b4-b5d8-342cc96b5b0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.761251] env[62183]: DEBUG oslo_vmware.api [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 960.761251] env[62183]: value = "task-1387304" [ 960.761251] env[62183]: _type = "Task" [ 960.761251] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.771763] env[62183]: DEBUG oslo_vmware.api [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387304, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.792348] env[62183]: DEBUG oslo_concurrency.lockutils [None req-122097e8-6c7a-4f43-88b7-693ebb856628 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "d30cfb69-1814-49b2-a29d-10cfbcc05722" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.166s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.907242] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.277s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.907912] env[62183]: DEBUG nova.compute.manager [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.910693] env[62183]: DEBUG oslo_vmware.api [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387301, 'name': RemoveSnapshot_Task, 'duration_secs': 0.806619} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.911244] env[62183]: DEBUG oslo_concurrency.lockutils [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.152s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.911455] env[62183]: DEBUG oslo_concurrency.lockutils [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.913559] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.510s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.915169] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deleted Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 960.915428] env[62183]: INFO nova.compute.manager [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Took 17.99 seconds to snapshot the instance on the hypervisor. [ 960.941776] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387302, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.950577] env[62183]: INFO nova.scheduler.client.report [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted allocations for instance 28d279bd-206e-4ba0-bf84-2980ab41f38d [ 961.235365] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387303, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.271507] env[62183]: DEBUG oslo_vmware.api [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387304, 'name': PowerOffVM_Task, 'duration_secs': 0.353512} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.271812] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 961.271991] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 961.272285] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95647d9c-d0e2-4b89-9913-708064308cd8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.361716] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 961.362064] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 961.362149] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleting the datastore file [datastore1] b89455d9-6d00-41ac-95ec-fb036b8c26c9 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.362412] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1aaea9c0-516e-45bf-9dd4-469424c11e99 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.370090] env[62183]: DEBUG oslo_vmware.api [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for the task: (returnval){ [ 961.370090] env[62183]: value = "task-1387306" [ 961.370090] env[62183]: _type = "Task" [ 961.370090] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.379346] env[62183]: DEBUG oslo_vmware.api [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387306, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.413225] env[62183]: DEBUG nova.compute.utils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 961.414632] env[62183]: DEBUG nova.compute.manager [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 961.414808] env[62183]: DEBUG nova.network.neutron [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 961.439916] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387302, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609032} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.440192] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 8c4b68d6-fedd-408f-a449-aace7400014a/8c4b68d6-fedd-408f-a449-aace7400014a.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.440406] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.440655] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-80b6a83e-51f5-4feb-8cd0-551c7df53075 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.447575] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 961.447575] env[62183]: value = "task-1387307" [ 961.447575] env[62183]: _type = "Task" [ 961.447575] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.455405] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387307, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.456910] env[62183]: DEBUG nova.policy [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '89ad04efd23b40e6a3eea5b3e6fd021e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b61c6a99cbda435481bb72f20929f03f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.463026] env[62183]: DEBUG oslo_concurrency.lockutils [None req-381bee32-6a70-4fcb-9244-1a9b55df33c2 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "28d279bd-206e-4ba0-bf84-2980ab41f38d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.181s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.472079] env[62183]: DEBUG nova.compute.manager [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Found 3 images (rotation: 2) {{(pid=62183) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 961.472294] env[62183]: DEBUG nova.compute.manager [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Rotating out 1 backups {{(pid=62183) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 961.472461] env[62183]: DEBUG nova.compute.manager [None req-e5565a11-c4fe-43ea-9688-145a89a842be tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deleting image aa118f5a-60fe-4272-b51b-366d31a6e666 {{(pid=62183) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 961.522096] env[62183]: DEBUG nova.network.neutron [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Updated VIF entry in instance network info cache for port 21c40087-5244-413a-ae18-e7e27f2abc34. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 961.522477] env[62183]: DEBUG nova.network.neutron [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Updating instance_info_cache with network_info: [{"id": "21c40087-5244-413a-ae18-e7e27f2abc34", "address": "fa:16:3e:92:b7:b9", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21c40087-52", "ovs_interfaceid": "21c40087-5244-413a-ae18-e7e27f2abc34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.606362] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.606631] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.606907] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.607088] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.607278] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.609647] env[62183]: INFO nova.compute.manager [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Terminating instance [ 961.611415] env[62183]: DEBUG nova.compute.manager [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.611614] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.612461] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6127f119-9531-4d3e-b223-71c694c99d99 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.621859] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.622277] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b90039ce-5a2d-4fe5-8d20-29cc26af5ba9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.628431] env[62183]: DEBUG oslo_vmware.api [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 961.628431] env[62183]: value = "task-1387308" [ 961.628431] env[62183]: _type = "Task" [ 961.628431] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.636183] env[62183]: DEBUG oslo_vmware.api [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387308, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.674367] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "d30cfb69-1814-49b2-a29d-10cfbcc05722" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.674687] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "d30cfb69-1814-49b2-a29d-10cfbcc05722" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.674952] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "d30cfb69-1814-49b2-a29d-10cfbcc05722-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.675175] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "d30cfb69-1814-49b2-a29d-10cfbcc05722-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.675350] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "d30cfb69-1814-49b2-a29d-10cfbcc05722-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.677936] env[62183]: INFO nova.compute.manager [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Terminating instance [ 961.682026] env[62183]: DEBUG nova.compute.manager [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.682026] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.682649] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca1f331-6d1c-4633-a1f5-dad90400d2ac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.690872] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.691150] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a839e31-4d95-47fa-96ea-7223df90e4e0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.698440] env[62183]: DEBUG oslo_vmware.api [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 961.698440] env[62183]: value = "task-1387309" [ 961.698440] env[62183]: _type = "Task" [ 961.698440] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.707199] env[62183]: DEBUG oslo_vmware.api [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387309, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.708112] env[62183]: DEBUG nova.network.neutron [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Successfully created port: 2678aa87-cdf5-4de0-b9f1-f00603f308cd {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.736723] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387303, 'name': CreateVM_Task, 'duration_secs': 0.72641} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.736966] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 961.737698] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.737952] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.738494] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 961.738671] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0313888-d879-4a09-8054-e3d340fcc368 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.743972] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 961.743972] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52ed3129-9cc2-aa3b-ddbd-b8d83964cd16" [ 961.743972] env[62183]: _type = "Task" [ 961.743972] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.753042] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ed3129-9cc2-aa3b-ddbd-b8d83964cd16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.879353] env[62183]: DEBUG oslo_vmware.api [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Task: {'id': task-1387306, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360056} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.879626] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.879814] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.879991] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.880191] env[62183]: INFO nova.compute.manager [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Took 1.14 seconds to destroy the instance on the hypervisor. [ 961.880437] env[62183]: DEBUG oslo.service.loopingcall [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.880648] env[62183]: DEBUG nova.compute.manager [-] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.880746] env[62183]: DEBUG nova.network.neutron [-] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 961.917411] env[62183]: DEBUG nova.compute.manager [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.953275] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.953522] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a948464b-63aa-4bc8-9885-228049e96d37 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954036] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954036] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954036] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance b89455d9-6d00-41ac-95ec-fb036b8c26c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954036] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance fb348784-62a0-4d1f-ac7f-f176f3da0dd9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954324] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954324] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance ecdd6f2a-80bd-495d-b54a-cf09ace67a87 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954450] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance d30cfb69-1814-49b2-a29d-10cfbcc05722 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954484] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 8c4b68d6-fedd-408f-a449-aace7400014a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954616] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 3a3ca703-2883-4aa9-a33f-326cc25d4838 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954683] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 961.954881] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 961.955746] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 961.961172] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387307, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071812} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.961576] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.962310] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d461fca-2984-4d58-ab7a-03b05e297b18 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.984978] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 8c4b68d6-fedd-408f-a449-aace7400014a/8c4b68d6-fedd-408f-a449-aace7400014a.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.988238] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8dba067e-609c-43a2-a257-baf68ca7def3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.008876] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 962.008876] env[62183]: value = "task-1387310" [ 962.008876] env[62183]: _type = "Task" [ 962.008876] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.018268] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387310, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.025362] env[62183]: DEBUG oslo_concurrency.lockutils [req-ec7fb929-5082-4374-82e9-b842f1c824e1 req-d6072865-0e6a-4653-9154-2f73b698c732 service nova] Releasing lock "refresh_cache-3a3ca703-2883-4aa9-a33f-326cc25d4838" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.140940] env[62183]: DEBUG oslo_vmware.api [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387308, 'name': PowerOffVM_Task, 'duration_secs': 0.287187} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.141237] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.141408] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.141661] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54c5e58d-002a-4337-bf88-3557c996a0fe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.162898] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b6acf1-d69b-4034-bd9f-031f4cec4aeb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.173600] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b182189-8834-44a0-8f07-f198099bb89a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.207178] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d6e39b-8d88-43ff-b412-5785b1c162b8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.213670] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.213911] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.214093] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleting the datastore file [datastore1] ecdd6f2a-80bd-495d-b54a-cf09ace67a87 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.214356] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8cd3ae3-923e-402f-b73e-cfe3d24f5b41 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.218955] env[62183]: DEBUG oslo_vmware.api [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387309, 'name': PowerOffVM_Task, 'duration_secs': 0.394331} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.221467] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.221646] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.221940] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-40f7ecfe-4dff-467b-935b-a0976ed842bc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.224409] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a10110-e23a-4d8b-8895-f1eae8ab5f74 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.229556] env[62183]: DEBUG oslo_vmware.api [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 962.229556] env[62183]: value = "task-1387312" [ 962.229556] env[62183]: _type = "Task" [ 962.229556] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.240620] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.247348] env[62183]: DEBUG oslo_vmware.api [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387312, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.257513] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ed3129-9cc2-aa3b-ddbd-b8d83964cd16, 'name': SearchDatastore_Task, 'duration_secs': 0.013772} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.257878] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.258161] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.258485] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.258641] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.258826] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.260043] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2d7965d-ffd9-40a5-81ef-b5ce9af2ed57 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.278170] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.278817] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.279724] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16c2867b-1dfc-4c71-8849-bc8f8a4b37e7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.285704] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 962.285704] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52694371-dc9f-931b-bfa3-0b9aa9d6dd75" [ 962.285704] env[62183]: _type = "Task" [ 962.285704] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.295810] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52694371-dc9f-931b-bfa3-0b9aa9d6dd75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.297032] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.297274] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.297572] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleting the datastore file [datastore1] d30cfb69-1814-49b2-a29d-10cfbcc05722 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.298730] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d14a553-bd08-4f77-b807-e8e510cbfa10 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.305489] env[62183]: DEBUG oslo_vmware.api [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 962.305489] env[62183]: value = "task-1387314" [ 962.305489] env[62183]: _type = "Task" [ 962.305489] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.318806] env[62183]: DEBUG oslo_vmware.api [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387314, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.519732] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387310, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.601255] env[62183]: DEBUG nova.network.neutron [-] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.713641] env[62183]: DEBUG nova.compute.manager [req-c5455e9f-9d74-4266-b4c1-9ba007d3e231 req-2bb0f020-275c-45c9-b31e-dd28b013d2e9 service nova] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Received event network-vif-deleted-b1af05d2-9221-4ac1-a4b4-82f7013cde5c {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.739583] env[62183]: DEBUG oslo_vmware.api [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387312, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.446341} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.740291] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.740291] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 962.740485] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 962.740667] env[62183]: INFO nova.compute.manager [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Took 1.13 seconds to destroy the instance on the hypervisor. [ 962.740912] env[62183]: DEBUG oslo.service.loopingcall [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.741130] env[62183]: DEBUG nova.compute.manager [-] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.741225] env[62183]: DEBUG nova.network.neutron [-] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 962.743543] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 962.796021] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52694371-dc9f-931b-bfa3-0b9aa9d6dd75, 'name': SearchDatastore_Task, 'duration_secs': 0.013953} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.796838] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a18a9d3c-ae77-4b4c-b7c0-e84c2c2d8911 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.801929] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 962.801929] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]529aabb3-685d-8fb9-4b45-e31d22ed56c1" [ 962.801929] env[62183]: _type = "Task" [ 962.801929] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.811911] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529aabb3-685d-8fb9-4b45-e31d22ed56c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.816677] env[62183]: DEBUG oslo_vmware.api [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387314, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.432515} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.816945] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.817326] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 962.817391] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 962.817515] env[62183]: INFO nova.compute.manager [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Took 1.14 seconds to destroy the instance on the hypervisor. [ 962.817751] env[62183]: DEBUG oslo.service.loopingcall [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.817941] env[62183]: DEBUG nova.compute.manager [-] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.818069] env[62183]: DEBUG nova.network.neutron [-] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 962.926433] env[62183]: DEBUG nova.compute.manager [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.952796] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.953291] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.953378] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.953600] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.953793] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.953963] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.954194] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.954399] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.954598] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.954784] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.955018] env[62183]: DEBUG nova.virt.hardware [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.956031] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a069e0ef-9437-4564-adff-58b0d8fca708 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.964443] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d837f7f8-efc8-4dcc-9975-6d544b4717be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.019631] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387310, 'name': ReconfigVM_Task, 'duration_secs': 0.770458} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.024030] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 8c4b68d6-fedd-408f-a449-aace7400014a/8c4b68d6-fedd-408f-a449-aace7400014a.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.024030] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e9b05f37-3809-453a-a529-afa1da2e6639 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.028460] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 963.028460] env[62183]: value = "task-1387315" [ 963.028460] env[62183]: _type = "Task" [ 963.028460] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.038064] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387315, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.103454] env[62183]: INFO nova.compute.manager [-] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Took 1.22 seconds to deallocate network for instance. [ 963.130575] env[62183]: DEBUG nova.compute.manager [req-7c446de3-5660-409e-9e0c-a4c286bce4bd req-5d9fe778-ae4e-45ae-9817-97e94de0f779 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Received event network-vif-deleted-9564d10c-89e8-4c70-b77d-b0c68416935d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.130575] env[62183]: INFO nova.compute.manager [req-7c446de3-5660-409e-9e0c-a4c286bce4bd req-5d9fe778-ae4e-45ae-9817-97e94de0f779 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Neutron deleted interface 9564d10c-89e8-4c70-b77d-b0c68416935d; detaching it from the instance and deleting it from the info cache [ 963.130575] env[62183]: DEBUG nova.network.neutron [req-7c446de3-5660-409e-9e0c-a4c286bce4bd req-5d9fe778-ae4e-45ae-9817-97e94de0f779 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.251865] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 963.252269] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.339s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.318090] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529aabb3-685d-8fb9-4b45-e31d22ed56c1, 'name': SearchDatastore_Task, 'duration_secs': 0.023231} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.318474] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.318913] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 3a3ca703-2883-4aa9-a33f-326cc25d4838/3a3ca703-2883-4aa9-a33f-326cc25d4838.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 963.319273] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-48b27adf-f141-4d0e-b1bc-2dbfd3003d3f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.325954] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 963.325954] env[62183]: value = "task-1387316" [ 963.325954] env[62183]: _type = "Task" [ 963.325954] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.334161] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387316, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.389817] env[62183]: DEBUG nova.network.neutron [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Successfully updated port: 2678aa87-cdf5-4de0-b9f1-f00603f308cd {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.522955] env[62183]: DEBUG nova.network.neutron [-] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.538775] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387315, 'name': Rename_Task, 'duration_secs': 0.320264} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.539094] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.540594] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b103463-d3f2-460c-b5c2-d2b1eeb39746 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.549626] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 963.549626] env[62183]: value = "task-1387317" [ 963.549626] env[62183]: _type = "Task" [ 963.549626] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.558937] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387317, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.606708] env[62183]: DEBUG nova.network.neutron [-] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.610845] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.611276] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.611925] env[62183]: DEBUG nova.objects.instance [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lazy-loading 'resources' on Instance uuid b89455d9-6d00-41ac-95ec-fb036b8c26c9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.631750] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dfd8c1c6-6b3f-4e8d-9668-64ea14f91721 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.641298] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a7cebb-eb67-44d6-868b-03c14dd9c04d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.668444] env[62183]: DEBUG nova.compute.manager [req-7c446de3-5660-409e-9e0c-a4c286bce4bd req-5d9fe778-ae4e-45ae-9817-97e94de0f779 service nova] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Detach interface failed, port_id=9564d10c-89e8-4c70-b77d-b0c68416935d, reason: Instance d30cfb69-1814-49b2-a29d-10cfbcc05722 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 963.836454] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387316, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.893221] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "refresh_cache-8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.893284] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "refresh_cache-8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.893526] env[62183]: DEBUG nova.network.neutron [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.026336] env[62183]: INFO nova.compute.manager [-] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Took 1.28 seconds to deallocate network for instance. [ 964.045617] env[62183]: DEBUG oslo_concurrency.lockutils [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "a948464b-63aa-4bc8-9885-228049e96d37" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.046050] env[62183]: DEBUG oslo_concurrency.lockutils [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.046286] env[62183]: DEBUG nova.compute.manager [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.047296] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b10fc6a-3744-4b64-8b44-dcffbb3bf258 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.060285] env[62183]: DEBUG nova.compute.manager [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62183) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 964.060880] env[62183]: DEBUG nova.objects.instance [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'flavor' on Instance uuid a948464b-63aa-4bc8-9885-228049e96d37 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.066411] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387317, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.110163] env[62183]: INFO nova.compute.manager [-] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Took 1.29 seconds to deallocate network for instance. [ 964.276331] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f06cc39-38c0-4e2d-aad7-f46eabf0cf54 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.284333] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b40daad-544c-4dca-b669-5ee89191752a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.316648] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0fffce-20be-4385-83fe-cd406dd75981 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.324179] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc90aa1-0342-4be3-84f3-5426258fdde0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.340113] env[62183]: DEBUG nova.compute.provider_tree [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.344390] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387316, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.869378} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.344750] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 3a3ca703-2883-4aa9-a33f-326cc25d4838/3a3ca703-2883-4aa9-a33f-326cc25d4838.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 964.344955] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.345211] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be7be609-d92c-434e-94ac-d1141b56d520 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.352136] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 964.352136] env[62183]: value = "task-1387318" [ 964.352136] env[62183]: _type = "Task" [ 964.352136] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.361404] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387318, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.424838] env[62183]: DEBUG nova.network.neutron [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 964.534026] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.551147] env[62183]: DEBUG nova.network.neutron [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Updating instance_info_cache with network_info: [{"id": "2678aa87-cdf5-4de0-b9f1-f00603f308cd", "address": "fa:16:3e:cf:5b:f9", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2678aa87-cd", "ovs_interfaceid": "2678aa87-cdf5-4de0-b9f1-f00603f308cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.562469] env[62183]: DEBUG oslo_vmware.api [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387317, 'name': PowerOnVM_Task, 'duration_secs': 0.876996} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.562882] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.563040] env[62183]: INFO nova.compute.manager [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Took 8.58 seconds to spawn the instance on the hypervisor. [ 964.563248] env[62183]: DEBUG nova.compute.manager [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.564045] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e28a25-0d41-4b4a-ac79-029663fe5d8b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.569341] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.569554] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3faab03d-0854-4347-afa2-f0cce7ea9e17 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.579337] env[62183]: DEBUG oslo_vmware.api [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 964.579337] env[62183]: value = "task-1387319" [ 964.579337] env[62183]: _type = "Task" [ 964.579337] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.586419] env[62183]: DEBUG oslo_vmware.api [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.619893] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.799155] env[62183]: DEBUG nova.compute.manager [req-f744e00c-f4b2-4ccf-9bfa-2b3c297a66ee req-9b357311-3481-43c2-8fc6-11532f90d71f service nova] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Received event network-vif-deleted-cea51b38-bd41-4f63-8713-bcc06c8cf941 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.846350] env[62183]: DEBUG nova.scheduler.client.report [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 964.860930] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387318, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074745} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.861715] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 964.862485] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59022049-daf3-4cbc-b38f-e3885522343d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.884565] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 3a3ca703-2883-4aa9-a33f-326cc25d4838/3a3ca703-2883-4aa9-a33f-326cc25d4838.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.885016] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fcf9a126-17c8-43ae-a13f-4ecf0a864ed8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.903567] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 964.903567] env[62183]: value = "task-1387320" [ 964.903567] env[62183]: _type = "Task" [ 964.903567] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.910921] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387320, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.054411] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "refresh_cache-8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.054657] env[62183]: DEBUG nova.compute.manager [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Instance network_info: |[{"id": "2678aa87-cdf5-4de0-b9f1-f00603f308cd", "address": "fa:16:3e:cf:5b:f9", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2678aa87-cd", "ovs_interfaceid": "2678aa87-cdf5-4de0-b9f1-f00603f308cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.055087] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:5b:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f1b507ed-cd2d-4c09-9d96-c47bde6a7774', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2678aa87-cdf5-4de0-b9f1-f00603f308cd', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.062579] env[62183]: DEBUG oslo.service.loopingcall [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.063126] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.063364] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a2540efc-d32a-4de6-99b8-c6a8e146e34c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.086145] env[62183]: INFO nova.compute.manager [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Took 20.85 seconds to build instance. [ 965.089207] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.089207] env[62183]: value = "task-1387321" [ 965.089207] env[62183]: _type = "Task" [ 965.089207] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.092336] env[62183]: DEBUG oslo_vmware.api [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387319, 'name': PowerOffVM_Task, 'duration_secs': 0.191016} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.095365] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.095555] env[62183]: DEBUG nova.compute.manager [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.096341] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f0ed1af-727d-454e-a1af-7da1906694c0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.104728] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387321, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.155895] env[62183]: DEBUG nova.compute.manager [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Received event network-vif-plugged-2678aa87-cdf5-4de0-b9f1-f00603f308cd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.156224] env[62183]: DEBUG oslo_concurrency.lockutils [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] Acquiring lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.156441] env[62183]: DEBUG oslo_concurrency.lockutils [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] Lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.156615] env[62183]: DEBUG oslo_concurrency.lockutils [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] Lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.156785] env[62183]: DEBUG nova.compute.manager [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] No waiting events found dispatching network-vif-plugged-2678aa87-cdf5-4de0-b9f1-f00603f308cd {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.157044] env[62183]: WARNING nova.compute.manager [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Received unexpected event network-vif-plugged-2678aa87-cdf5-4de0-b9f1-f00603f308cd for instance with vm_state building and task_state spawning. [ 965.157243] env[62183]: DEBUG nova.compute.manager [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Received event network-changed-2678aa87-cdf5-4de0-b9f1-f00603f308cd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.157405] env[62183]: DEBUG nova.compute.manager [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Refreshing instance network info cache due to event network-changed-2678aa87-cdf5-4de0-b9f1-f00603f308cd. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.157590] env[62183]: DEBUG oslo_concurrency.lockutils [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] Acquiring lock "refresh_cache-8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.157728] env[62183]: DEBUG oslo_concurrency.lockutils [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] Acquired lock "refresh_cache-8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.157885] env[62183]: DEBUG nova.network.neutron [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Refreshing network info cache for port 2678aa87-cdf5-4de0-b9f1-f00603f308cd {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.350763] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.739s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.353175] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.819s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.353452] env[62183]: DEBUG nova.objects.instance [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lazy-loading 'resources' on Instance uuid ecdd6f2a-80bd-495d-b54a-cf09ace67a87 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.374885] env[62183]: INFO nova.scheduler.client.report [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Deleted allocations for instance b89455d9-6d00-41ac-95ec-fb036b8c26c9 [ 965.415888] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387320, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.588369] env[62183]: DEBUG oslo_concurrency.lockutils [None req-655fbcd6-eb95-48c4-9739-75f03617fc6d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.362s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.602616] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387321, 'name': CreateVM_Task, 'duration_secs': 0.408574} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.604075] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 965.604075] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.604075] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.604335] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 965.604606] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ecc0405-435d-4b81-80ce-4e158d3ccf79 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.611861] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 965.611861] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a23884-241e-4494-3914-29dc8a51cbd5" [ 965.611861] env[62183]: _type = "Task" [ 965.611861] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.612401] env[62183]: DEBUG oslo_concurrency.lockutils [None req-eeb520cd-8a38-4cbb-a98e-320dd8d77b25 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.566s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.621908] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a23884-241e-4494-3914-29dc8a51cbd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.883153] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6b7e0108-1586-4da6-b6a6-2a5819000eb0 tempest-DeleteServersTestJSON-105649048 tempest-DeleteServersTestJSON-105649048-project-member] Lock "b89455d9-6d00-41ac-95ec-fb036b8c26c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.144s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.918852] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387320, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.986026] env[62183]: DEBUG nova.network.neutron [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Updated VIF entry in instance network info cache for port 2678aa87-cdf5-4de0-b9f1-f00603f308cd. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.986866] env[62183]: DEBUG nova.network.neutron [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Updating instance_info_cache with network_info: [{"id": "2678aa87-cdf5-4de0-b9f1-f00603f308cd", "address": "fa:16:3e:cf:5b:f9", "network": {"id": "d6ea7337-1dc6-4bba-bfee-52911f25b5bd", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-615373888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b61c6a99cbda435481bb72f20929f03f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f1b507ed-cd2d-4c09-9d96-c47bde6a7774", "external-id": "nsx-vlan-transportzone-980", "segmentation_id": 980, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2678aa87-cd", "ovs_interfaceid": "2678aa87-cdf5-4de0-b9f1-f00603f308cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.033420] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d99bc9c-8529-462d-be17-bbd0fa3ce2e5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.042089] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3466d99-835b-45c1-bc51-5ed5426bf79d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.077776] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883babe2-f634-4015-b6c3-dd207413426b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.086467] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60962a17-a3d4-4000-9d81-21113df730a9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.101571] env[62183]: DEBUG nova.compute.provider_tree [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.122673] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a23884-241e-4494-3914-29dc8a51cbd5, 'name': SearchDatastore_Task, 'duration_secs': 0.069268} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.123016] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.123288] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.123531] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.123682] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.123879] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.124165] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-11245e73-2404-43b9-b635-497a723917b2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.133647] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.133872] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 966.134596] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a30ddb3c-3c66-475a-ae7d-cc2650e800b8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.140502] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 966.140502] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52134a8c-bf98-24c6-6c8f-6636e33c50f5" [ 966.140502] env[62183]: _type = "Task" [ 966.140502] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.150425] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52134a8c-bf98-24c6-6c8f-6636e33c50f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.418466] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387320, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.493334] env[62183]: DEBUG oslo_concurrency.lockutils [req-b2256861-bde2-42ff-8a80-c9fd880c06b0 req-913aa9e0-31b8-4187-8f6b-59fa0ed0d015 service nova] Releasing lock "refresh_cache-8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.606462] env[62183]: DEBUG nova.scheduler.client.report [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 966.651548] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52134a8c-bf98-24c6-6c8f-6636e33c50f5, 'name': SearchDatastore_Task, 'duration_secs': 0.01064} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.653032] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83066019-36b3-48f1-92f0-2079e071c5b6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.659050] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 966.659050] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]521a2ed6-0972-5b12-97a5-92db65022581" [ 966.659050] env[62183]: _type = "Task" [ 966.659050] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.667802] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521a2ed6-0972-5b12-97a5-92db65022581, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.828245] env[62183]: DEBUG nova.compute.manager [req-5f1737e7-38e2-43eb-8839-e4e9bb0cc4a4 req-c947f54b-74f9-454a-a96c-3cbcdd9801eb service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Received event network-changed-e301232c-2b72-4bc5-b5f5-811b208bcc94 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.828569] env[62183]: DEBUG nova.compute.manager [req-5f1737e7-38e2-43eb-8839-e4e9bb0cc4a4 req-c947f54b-74f9-454a-a96c-3cbcdd9801eb service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Refreshing instance network info cache due to event network-changed-e301232c-2b72-4bc5-b5f5-811b208bcc94. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 966.828928] env[62183]: DEBUG oslo_concurrency.lockutils [req-5f1737e7-38e2-43eb-8839-e4e9bb0cc4a4 req-c947f54b-74f9-454a-a96c-3cbcdd9801eb service nova] Acquiring lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.829197] env[62183]: DEBUG oslo_concurrency.lockutils [req-5f1737e7-38e2-43eb-8839-e4e9bb0cc4a4 req-c947f54b-74f9-454a-a96c-3cbcdd9801eb service nova] Acquired lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.829474] env[62183]: DEBUG nova.network.neutron [req-5f1737e7-38e2-43eb-8839-e4e9bb0cc4a4 req-c947f54b-74f9-454a-a96c-3cbcdd9801eb service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Refreshing network info cache for port e301232c-2b72-4bc5-b5f5-811b208bcc94 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 966.920664] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387320, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.115100] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.117612] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.498s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.117906] env[62183]: DEBUG nova.objects.instance [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lazy-loading 'resources' on Instance uuid d30cfb69-1814-49b2-a29d-10cfbcc05722 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.141967] env[62183]: INFO nova.scheduler.client.report [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleted allocations for instance ecdd6f2a-80bd-495d-b54a-cf09ace67a87 [ 967.172625] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521a2ed6-0972-5b12-97a5-92db65022581, 'name': SearchDatastore_Task, 'duration_secs': 0.011053} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.172893] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.173274] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec/8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 967.173597] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9227fd71-0ec5-4296-9091-b8420e3ced86 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.183690] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 967.183690] env[62183]: value = "task-1387322" [ 967.183690] env[62183]: _type = "Task" [ 967.183690] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.192387] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387322, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.383351] env[62183]: DEBUG nova.compute.manager [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Stashing vm_state: stopped {{(pid=62183) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 967.435432] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387320, 'name': ReconfigVM_Task, 'duration_secs': 2.017802} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.436220] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 3a3ca703-2883-4aa9-a33f-326cc25d4838/3a3ca703-2883-4aa9-a33f-326cc25d4838.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.439254] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0875b0d8-73e7-4eb4-96e8-2fcdf7f8b74e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.451022] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 967.451022] env[62183]: value = "task-1387323" [ 967.451022] env[62183]: _type = "Task" [ 967.451022] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.471277] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387323, 'name': Rename_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.658083] env[62183]: DEBUG oslo_concurrency.lockutils [None req-cc845e0a-b1e6-4207-a1a7-81970d9dd68d tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "ecdd6f2a-80bd-495d-b54a-cf09ace67a87" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.051s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.699221] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387322, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.824981] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0d52b4-c56b-4d2d-b819-a9b96f73d8aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.830353] env[62183]: DEBUG nova.network.neutron [req-5f1737e7-38e2-43eb-8839-e4e9bb0cc4a4 req-c947f54b-74f9-454a-a96c-3cbcdd9801eb service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updated VIF entry in instance network info cache for port e301232c-2b72-4bc5-b5f5-811b208bcc94. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 967.830816] env[62183]: DEBUG nova.network.neutron [req-5f1737e7-38e2-43eb-8839-e4e9bb0cc4a4 req-c947f54b-74f9-454a-a96c-3cbcdd9801eb service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updating instance_info_cache with network_info: [{"id": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "address": "fa:16:3e:10:9e:d2", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape301232c-2b", "ovs_interfaceid": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.835541] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a757570b-cf79-4469-825d-c10a7dec5846 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.869117] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729b00c0-9f0a-4dca-9a66-d4a69533be59 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.877524] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000d04b9-613d-4824-8851-b37a0e051ccf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.896472] env[62183]: DEBUG nova.compute.provider_tree [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.909844] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.963603] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387323, 'name': Rename_Task, 'duration_secs': 0.389893} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.963935] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 967.964225] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-943ff865-1350-4b80-9fd9-d38d1c7e824e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.972326] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 967.972326] env[62183]: value = "task-1387324" [ 967.972326] env[62183]: _type = "Task" [ 967.972326] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.980258] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387324, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.200362] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387322, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562633} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.200744] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec/8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.201071] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.201418] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fe02d1ae-c500-4ea8-bc27-81b7d348f354 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.211070] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 968.211070] env[62183]: value = "task-1387325" [ 968.211070] env[62183]: _type = "Task" [ 968.211070] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.218952] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387325, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.336163] env[62183]: DEBUG oslo_concurrency.lockutils [req-5f1737e7-38e2-43eb-8839-e4e9bb0cc4a4 req-c947f54b-74f9-454a-a96c-3cbcdd9801eb service nova] Releasing lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.400283] env[62183]: DEBUG nova.scheduler.client.report [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 968.483694] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387324, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.720975] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387325, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074527} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.721281] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.722085] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d62f08d-05a0-4f39-b580-3a5ec11590d4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.745920] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec/8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.746256] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ec46542-39ce-4b0a-a90c-265417606189 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.773226] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 968.773226] env[62183]: value = "task-1387326" [ 968.773226] env[62183]: _type = "Task" [ 968.773226] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.785283] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387326, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.904808] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.787s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.908031] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.997s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.931080] env[62183]: INFO nova.scheduler.client.report [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleted allocations for instance d30cfb69-1814-49b2-a29d-10cfbcc05722 [ 968.985067] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387324, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.283302] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387326, 'name': ReconfigVM_Task, 'duration_secs': 0.336625} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.283674] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec/8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.284349] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73245472-dbfd-4f9b-a0c9-6e4f0d4cd7e0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.292164] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 969.292164] env[62183]: value = "task-1387327" [ 969.292164] env[62183]: _type = "Task" [ 969.292164] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.300952] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387327, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.411912] env[62183]: INFO nova.compute.claims [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.441053] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9096d7c6-9dce-43ef-8321-7d9acb72d75a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "d30cfb69-1814-49b2-a29d-10cfbcc05722" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.766s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.485658] env[62183]: DEBUG oslo_vmware.api [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387324, 'name': PowerOnVM_Task, 'duration_secs': 1.016118} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.485932] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 969.486217] env[62183]: INFO nova.compute.manager [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Took 11.13 seconds to spawn the instance on the hypervisor. [ 969.486415] env[62183]: DEBUG nova.compute.manager [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 969.487381] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa0cb6e-ebca-46ca-aabf-8114aa14b458 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.803833] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387327, 'name': Rename_Task, 'duration_secs': 0.149338} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.808954] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.809227] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30ab099b-06f6-434b-8b2a-98c54cfea85c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.820454] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 969.820454] env[62183]: value = "task-1387328" [ 969.820454] env[62183]: _type = "Task" [ 969.820454] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.830181] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387328, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.918713] env[62183]: INFO nova.compute.resource_tracker [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating resource usage from migration 4e58e303-0d2b-4df3-96ab-52a8e649d12b [ 970.009079] env[62183]: INFO nova.compute.manager [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Took 24.22 seconds to build instance. [ 970.103152] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e6f1e9-94fe-4e12-8e70-a4178565f14c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.112025] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f65bb8-db4d-462e-8d5a-2e38283da841 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.144659] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebed2f02-1d65-4c8d-ba39-6788a4cb7590 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.153094] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0f7948-dd8a-4d6d-aa6b-7a8b3fb7c003 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.170057] env[62183]: DEBUG nova.compute.provider_tree [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.328948] env[62183]: DEBUG oslo_vmware.api [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387328, 'name': PowerOnVM_Task, 'duration_secs': 0.497303} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.329639] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.329974] env[62183]: INFO nova.compute.manager [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Took 7.40 seconds to spawn the instance on the hypervisor. [ 970.332021] env[62183]: DEBUG nova.compute.manager [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.332021] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13ca660-5055-4b53-b4ba-b788e7ee290b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.512492] env[62183]: DEBUG oslo_concurrency.lockutils [None req-28713660-0244-4b01-9552-13b511b27fc8 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "3a3ca703-2883-4aa9-a33f-326cc25d4838" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.736s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.673703] env[62183]: DEBUG nova.scheduler.client.report [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 970.851021] env[62183]: INFO nova.compute.manager [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Took 16.99 seconds to build instance. [ 970.985134] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.985417] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.016606] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "143323db-da1c-4dd8-8f13-ed0af31e2027" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.016879] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "143323db-da1c-4dd8-8f13-ed0af31e2027" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.179924] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.273s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.180212] env[62183]: INFO nova.compute.manager [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Migrating [ 971.238085] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquiring lock "036037b3-9676-419c-ab87-e5a557438fbe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.238418] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "036037b3-9676-419c-ab87-e5a557438fbe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.353457] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c64a79e3-46e6-4b04-b745-5f2f9cbeb119 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.500s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.488388] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.519698] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.582425] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.582634] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.694452] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.694806] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.694963] env[62183]: DEBUG nova.network.neutron [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 971.741185] env[62183]: DEBUG nova.compute.manager [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.017394] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.017394] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.019135] env[62183]: INFO nova.compute.claims [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.043689] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.084749] env[62183]: DEBUG nova.compute.manager [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.260938] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.508124] env[62183]: DEBUG nova.network.neutron [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [{"id": "3d687443-41fd-44ff-8981-35319b79d88a", "address": "fa:16:3e:91:31:9b", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d687443-41", "ovs_interfaceid": "3d687443-41fd-44ff-8981-35319b79d88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.612471] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.962447] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.962733] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.962956] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.963163] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.963370] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.965711] env[62183]: INFO nova.compute.manager [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Terminating instance [ 972.967650] env[62183]: DEBUG nova.compute.manager [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.967847] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 972.968746] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa17f213-54ec-40e9-ba86-950a059b417b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.977740] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.977942] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a38104c-394d-4f54-b45b-43d7fca4f7cb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.985343] env[62183]: DEBUG oslo_vmware.api [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 972.985343] env[62183]: value = "task-1387329" [ 972.985343] env[62183]: _type = "Task" [ 972.985343] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.994144] env[62183]: DEBUG oslo_vmware.api [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.009934] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.219945] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7a2e9a-7b3a-43e9-a5a1-cc83ed7fbe84 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.228751] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e86fe0-d973-4b94-9652-68b3c2144906 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.259419] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f937bd5e-42a1-42a1-92b2-6f7874d2a0eb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.268033] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ae50a9-1127-4fef-8be2-dd6b5fbbf990 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.282163] env[62183]: DEBUG nova.compute.provider_tree [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.342131] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "d8817d24-b0cb-4956-b195-cc417ae09fb4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.342364] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "d8817d24-b0cb-4956-b195-cc417ae09fb4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.445608] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.445846] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.497170] env[62183]: DEBUG oslo_vmware.api [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387329, 'name': PowerOffVM_Task, 'duration_secs': 0.20943} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.497638] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.497983] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 973.498375] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-809d22ca-5be7-467d-a924-8e2ff96f3528 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.565489] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 973.565689] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 973.565883] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleting the datastore file [datastore1] 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.566277] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff7c2590-5098-4132-b869-da262a908858 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.574172] env[62183]: DEBUG oslo_vmware.api [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for the task: (returnval){ [ 973.574172] env[62183]: value = "task-1387331" [ 973.574172] env[62183]: _type = "Task" [ 973.574172] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.584559] env[62183]: DEBUG oslo_vmware.api [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387331, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.785881] env[62183]: DEBUG nova.scheduler.client.report [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 973.844271] env[62183]: DEBUG nova.compute.manager [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 973.952189] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 973.952401] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 974.085268] env[62183]: DEBUG oslo_vmware.api [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Task: {'id': task-1387331, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.503347} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.085506] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.085692] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 974.085870] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 974.086057] env[62183]: INFO nova.compute.manager [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Took 1.12 seconds to destroy the instance on the hypervisor. [ 974.086391] env[62183]: DEBUG oslo.service.loopingcall [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.086597] env[62183]: DEBUG nova.compute.manager [-] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 974.086694] env[62183]: DEBUG nova.network.neutron [-] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 974.291314] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.291999] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.295742] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.253s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.297367] env[62183]: INFO nova.compute.claims [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.367350] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.401144] env[62183]: DEBUG nova.compute.manager [req-262b3984-ddea-4518-bdcf-368ad5e4c06e req-8b2fb4ac-e819-4e55-b66f-75aa8d9644b9 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Received event network-vif-deleted-2678aa87-cdf5-4de0-b9f1-f00603f308cd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.401368] env[62183]: INFO nova.compute.manager [req-262b3984-ddea-4518-bdcf-368ad5e4c06e req-8b2fb4ac-e819-4e55-b66f-75aa8d9644b9 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Neutron deleted interface 2678aa87-cdf5-4de0-b9f1-f00603f308cd; detaching it from the instance and deleting it from the info cache [ 974.401550] env[62183]: DEBUG nova.network.neutron [req-262b3984-ddea-4518-bdcf-368ad5e4c06e req-8b2fb4ac-e819-4e55-b66f-75aa8d9644b9 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.525732] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b809863-0624-4535-9342-295ebec7d316 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.545531] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance 'a948464b-63aa-4bc8-9885-228049e96d37' progress to 0 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 974.802559] env[62183]: DEBUG nova.compute.utils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.805699] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 974.805874] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 974.851124] env[62183]: DEBUG nova.policy [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b92c15df0ab4e7f9e845e97d3e5912c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '499aba16038147e3b20b06876234cb13', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.878610] env[62183]: DEBUG nova.network.neutron [-] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.904420] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87a092de-7c8a-43f6-9728-9e4488a531bc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.916641] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b103edfa-756f-4238-8c74-6b4b07349c62 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.947806] env[62183]: DEBUG nova.compute.manager [req-262b3984-ddea-4518-bdcf-368ad5e4c06e req-8b2fb4ac-e819-4e55-b66f-75aa8d9644b9 service nova] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Detach interface failed, port_id=2678aa87-cdf5-4de0-b9f1-f00603f308cd, reason: Instance 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 974.957565] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.957825] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquired lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.957881] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Forcefully refreshing network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 975.051425] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 975.051748] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f3db03a-c4aa-4571-b18f-d879c8cdf596 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.061283] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 975.061283] env[62183]: value = "task-1387332" [ 975.061283] env[62183]: _type = "Task" [ 975.061283] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.071677] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] VM already powered off {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 975.071914] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance 'a948464b-63aa-4bc8-9885-228049e96d37' progress to 17 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 975.114082] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Successfully created port: 9cc46a34-db54-4e5b-97fe-e3b4968da907 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.306604] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 975.382851] env[62183]: INFO nova.compute.manager [-] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Took 1.30 seconds to deallocate network for instance. [ 975.498447] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2eb0e7-9ed8-4745-a353-c4795e7130ea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.508460] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ffda07-a3c1-40fd-94cf-8da5e956144e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.541950] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605920e2-bae9-48f4-843a-59cf33eae33d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.550233] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326b12fb-fb07-4576-898b-9a700c8ccaa5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.563900] env[62183]: DEBUG nova.compute.provider_tree [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.577835] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.578076] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.578241] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.578427] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.578577] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.578723] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.578927] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.579109] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.579363] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.579560] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.579738] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.584923] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9bbdecc-226b-4730-971d-9bedd6a625ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.601580] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 975.601580] env[62183]: value = "task-1387333" [ 975.601580] env[62183]: _type = "Task" [ 975.601580] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.610346] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.889195] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.066660] env[62183]: DEBUG nova.scheduler.client.report [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 976.112289] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387333, 'name': ReconfigVM_Task, 'duration_secs': 0.147401} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.112598] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance 'a948464b-63aa-4bc8-9885-228049e96d37' progress to 33 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 976.175917] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [{"id": "3d687443-41fd-44ff-8981-35319b79d88a", "address": "fa:16:3e:91:31:9b", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d687443-41", "ovs_interfaceid": "3d687443-41fd-44ff-8981-35319b79d88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.318974] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.348247] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.348519] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.348680] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.348862] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.349015] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.349171] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.349382] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.349542] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.349712] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.349876] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.350060] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.350919] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99bca829-3ffc-49de-b083-2c6b6da1635f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.360350] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c00f67de-9398-48df-8cf2-70de8df684ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.524471] env[62183]: DEBUG nova.compute.manager [req-dfa52e16-5710-4b38-9038-05e8aa4936ac req-69433cf5-d491-43d7-82c6-55d7dacb43c9 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Received event network-vif-plugged-9cc46a34-db54-4e5b-97fe-e3b4968da907 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.524784] env[62183]: DEBUG oslo_concurrency.lockutils [req-dfa52e16-5710-4b38-9038-05e8aa4936ac req-69433cf5-d491-43d7-82c6-55d7dacb43c9 service nova] Acquiring lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.525052] env[62183]: DEBUG oslo_concurrency.lockutils [req-dfa52e16-5710-4b38-9038-05e8aa4936ac req-69433cf5-d491-43d7-82c6-55d7dacb43c9 service nova] Lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.525236] env[62183]: DEBUG oslo_concurrency.lockutils [req-dfa52e16-5710-4b38-9038-05e8aa4936ac req-69433cf5-d491-43d7-82c6-55d7dacb43c9 service nova] Lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.525412] env[62183]: DEBUG nova.compute.manager [req-dfa52e16-5710-4b38-9038-05e8aa4936ac req-69433cf5-d491-43d7-82c6-55d7dacb43c9 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] No waiting events found dispatching network-vif-plugged-9cc46a34-db54-4e5b-97fe-e3b4968da907 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.525889] env[62183]: WARNING nova.compute.manager [req-dfa52e16-5710-4b38-9038-05e8aa4936ac req-69433cf5-d491-43d7-82c6-55d7dacb43c9 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Received unexpected event network-vif-plugged-9cc46a34-db54-4e5b-97fe-e3b4968da907 for instance with vm_state building and task_state spawning. [ 976.571466] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.572016] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 976.574620] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.314s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.577842] env[62183]: INFO nova.compute.claims [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.619568] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.619568] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.619772] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.621020] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.621020] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.621020] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.621020] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.621020] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.621020] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.621020] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.621020] env[62183]: DEBUG nova.virt.hardware [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.626992] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Reconfiguring VM instance instance-00000048 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 976.627558] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4271beb5-90c9-4476-b05d-a10138e39e2a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.647996] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 976.647996] env[62183]: value = "task-1387334" [ 976.647996] env[62183]: _type = "Task" [ 976.647996] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.657033] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387334, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.666830] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Successfully updated port: 9cc46a34-db54-4e5b-97fe-e3b4968da907 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 976.678588] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Releasing lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.678819] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updated the network info_cache for instance {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 976.679652] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.679852] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.680012] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.680174] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.680332] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.680487] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.680640] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 976.681181] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 977.081632] env[62183]: DEBUG nova.compute.utils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.085047] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 977.085047] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 977.126894] env[62183]: DEBUG nova.policy [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b92c15df0ab4e7f9e845e97d3e5912c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '499aba16038147e3b20b06876234cb13', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 977.158829] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387334, 'name': ReconfigVM_Task, 'duration_secs': 0.167495} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.159132] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Reconfigured VM instance instance-00000048 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 977.159960] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5801f25-fd49-42bc-968d-44946785bd8d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.177468] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "refresh_cache-b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.177620] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "refresh_cache-b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.177772] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 977.188386] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] a948464b-63aa-4bc8-9885-228049e96d37/a948464b-63aa-4bc8-9885-228049e96d37.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.191774] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.192047] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88bb8161-c9bc-42cd-a470-dc17818008c2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.213114] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 977.213114] env[62183]: value = "task-1387335" [ 977.213114] env[62183]: _type = "Task" [ 977.213114] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.222837] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387335, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.249591] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 977.481071] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Updating instance_info_cache with network_info: [{"id": "9cc46a34-db54-4e5b-97fe-e3b4968da907", "address": "fa:16:3e:f4:25:df", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cc46a34-db", "ovs_interfaceid": "9cc46a34-db54-4e5b-97fe-e3b4968da907", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.522751] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Successfully created port: 81a4f8ba-127a-4d45-9fe2-462656e5ff9d {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 977.586162] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 977.724820] env[62183]: DEBUG oslo_vmware.api [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387335, 'name': ReconfigVM_Task, 'duration_secs': 0.308991} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.727389] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Reconfigured VM instance instance-00000048 to attach disk [datastore2] a948464b-63aa-4bc8-9885-228049e96d37/a948464b-63aa-4bc8-9885-228049e96d37.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.727667] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance 'a948464b-63aa-4bc8-9885-228049e96d37' progress to 50 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 977.788100] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d60ce0-fcfe-4ebf-b239-a4e6fc7d1236 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.797781] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b53f3d-e26f-4bcf-9e96-7ecec0bd28e0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.827509] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352fe8f7-8dc6-4bdb-8bc7-57c17fc841cb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.835259] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4940e802-4330-4a2c-89bf-91a019791e8d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.848818] env[62183]: DEBUG nova.compute.provider_tree [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.983354] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "refresh_cache-b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.983691] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Instance network_info: |[{"id": "9cc46a34-db54-4e5b-97fe-e3b4968da907", "address": "fa:16:3e:f4:25:df", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cc46a34-db", "ovs_interfaceid": "9cc46a34-db54-4e5b-97fe-e3b4968da907", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 977.984167] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:25:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4df917f7-847a-4c0e-b0e3-69a52e4a1554', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9cc46a34-db54-4e5b-97fe-e3b4968da907', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.991853] env[62183]: DEBUG oslo.service.loopingcall [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.992428] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.992667] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73d7f348-ff78-492b-9c1b-74556f2882ba {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.013401] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.013401] env[62183]: value = "task-1387336" [ 978.013401] env[62183]: _type = "Task" [ 978.013401] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.021521] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387336, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.234165] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3488a94d-16ca-4bb2-9120-64d583c15078 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.256545] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec686aed-fb88-43bb-8a47-d05a3578075e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.275464] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance 'a948464b-63aa-4bc8-9885-228049e96d37' progress to 67 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 978.351746] env[62183]: DEBUG nova.scheduler.client.report [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 978.523530] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387336, 'name': CreateVM_Task, 'duration_secs': 0.397365} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.523864] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 978.524446] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.524625] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.524950] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 978.525222] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-562626f6-4321-4dbf-b8b3-a4fd9b517a4d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.530272] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 978.530272] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e10998-a3cd-2ff1-4721-8898edee705d" [ 978.530272] env[62183]: _type = "Task" [ 978.530272] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.540610] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e10998-a3cd-2ff1-4721-8898edee705d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.551077] env[62183]: DEBUG nova.compute.manager [req-f2ebd9e1-79e1-47f5-8e1b-19742429fb38 req-f88e8096-a5a0-40fe-924a-fbfc1f557ec9 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Received event network-changed-9cc46a34-db54-4e5b-97fe-e3b4968da907 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.551287] env[62183]: DEBUG nova.compute.manager [req-f2ebd9e1-79e1-47f5-8e1b-19742429fb38 req-f88e8096-a5a0-40fe-924a-fbfc1f557ec9 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Refreshing instance network info cache due to event network-changed-9cc46a34-db54-4e5b-97fe-e3b4968da907. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.551498] env[62183]: DEBUG oslo_concurrency.lockutils [req-f2ebd9e1-79e1-47f5-8e1b-19742429fb38 req-f88e8096-a5a0-40fe-924a-fbfc1f557ec9 service nova] Acquiring lock "refresh_cache-b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.551677] env[62183]: DEBUG oslo_concurrency.lockutils [req-f2ebd9e1-79e1-47f5-8e1b-19742429fb38 req-f88e8096-a5a0-40fe-924a-fbfc1f557ec9 service nova] Acquired lock "refresh_cache-b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.551858] env[62183]: DEBUG nova.network.neutron [req-f2ebd9e1-79e1-47f5-8e1b-19742429fb38 req-f88e8096-a5a0-40fe-924a-fbfc1f557ec9 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Refreshing network info cache for port 9cc46a34-db54-4e5b-97fe-e3b4968da907 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.598773] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 978.623713] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 978.623974] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 978.624163] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 978.624354] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 978.624504] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 978.624657] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 978.624872] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 978.625041] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 978.625216] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 978.625402] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 978.625596] env[62183]: DEBUG nova.virt.hardware [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.626501] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57a5ae1-ac4f-4868-84d7-263fae8bc345 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.635679] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e342e3-d6a6-46e2-8d13-8de115e554cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.824668] env[62183]: DEBUG nova.network.neutron [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Port 3d687443-41fd-44ff-8981-35319b79d88a binding to destination host cpu-1 is already ACTIVE {{(pid=62183) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 978.857206] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.857766] env[62183]: DEBUG nova.compute.manager [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.860206] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.248s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.861983] env[62183]: INFO nova.compute.claims [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.042065] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e10998-a3cd-2ff1-4721-8898edee705d, 'name': SearchDatastore_Task, 'duration_secs': 0.032525} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.042433] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.042710] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.043030] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.043226] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.043447] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.043737] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-937864dd-bc30-4c33-a798-64cea3852233 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.053327] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.053990] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.056805] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-188d3735-c4fd-43ff-8c67-842f6dfbbd30 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.062620] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 979.062620] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5258e3ff-ca7a-9144-f385-63bc20e6f37b" [ 979.062620] env[62183]: _type = "Task" [ 979.062620] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.072407] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5258e3ff-ca7a-9144-f385-63bc20e6f37b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.138789] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Successfully updated port: 81a4f8ba-127a-4d45-9fe2-462656e5ff9d {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.350579] env[62183]: DEBUG nova.network.neutron [req-f2ebd9e1-79e1-47f5-8e1b-19742429fb38 req-f88e8096-a5a0-40fe-924a-fbfc1f557ec9 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Updated VIF entry in instance network info cache for port 9cc46a34-db54-4e5b-97fe-e3b4968da907. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 979.350939] env[62183]: DEBUG nova.network.neutron [req-f2ebd9e1-79e1-47f5-8e1b-19742429fb38 req-f88e8096-a5a0-40fe-924a-fbfc1f557ec9 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Updating instance_info_cache with network_info: [{"id": "9cc46a34-db54-4e5b-97fe-e3b4968da907", "address": "fa:16:3e:f4:25:df", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cc46a34-db", "ovs_interfaceid": "9cc46a34-db54-4e5b-97fe-e3b4968da907", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.365923] env[62183]: DEBUG nova.compute.utils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.369088] env[62183]: DEBUG nova.compute.manager [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 979.369277] env[62183]: DEBUG nova.network.neutron [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.406784] env[62183]: DEBUG nova.policy [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1589c8afb2ca44b9a30b15f35bee6cf7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '218d7143fba34b8081e826b83968d53d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 979.574944] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5258e3ff-ca7a-9144-f385-63bc20e6f37b, 'name': SearchDatastore_Task, 'duration_secs': 0.009784} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.574944] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5aee4fb7-38bd-4641-aa5b-8984b2709aaf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.580021] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 979.580021] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a64ef2-2985-fbcf-fd9e-d8c89551c901" [ 979.580021] env[62183]: _type = "Task" [ 979.580021] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.587676] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a64ef2-2985-fbcf-fd9e-d8c89551c901, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.642019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "refresh_cache-143323db-da1c-4dd8-8f13-ed0af31e2027" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.642190] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "refresh_cache-143323db-da1c-4dd8-8f13-ed0af31e2027" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.642349] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 979.644225] env[62183]: DEBUG nova.network.neutron [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Successfully created port: f9702809-bb25-463e-a193-580acaeaf306 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.847573] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "a948464b-63aa-4bc8-9885-228049e96d37-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.847875] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.847964] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.853066] env[62183]: DEBUG oslo_concurrency.lockutils [req-f2ebd9e1-79e1-47f5-8e1b-19742429fb38 req-f88e8096-a5a0-40fe-924a-fbfc1f557ec9 service nova] Releasing lock "refresh_cache-b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.869845] env[62183]: DEBUG nova.compute.manager [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 980.050039] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c414eb0-4976-4930-b538-edc12eff5016 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.058977] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f38eee1-6b3d-48ff-a0a2-c8d44ab47645 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.094567] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3069fd-4a5a-43e3-ab22-78bd758a66fb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.106281] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01e458a-f73a-42e9-9ea1-d7584c29fe43 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.110053] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a64ef2-2985-fbcf-fd9e-d8c89551c901, 'name': SearchDatastore_Task, 'duration_secs': 0.01339} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.110658] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.110923] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] b6402e2f-1035-4c1b-9a1e-05b17d89e4bb/b6402e2f-1035-4c1b-9a1e-05b17d89e4bb.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.111575] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2e22c3a-6a0c-4f29-90f1-b1592a0f4556 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.121134] env[62183]: DEBUG nova.compute.provider_tree [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.128613] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 980.128613] env[62183]: value = "task-1387337" [ 980.128613] env[62183]: _type = "Task" [ 980.128613] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.138190] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.187697] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 980.329976] env[62183]: DEBUG nova.network.neutron [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Updating instance_info_cache with network_info: [{"id": "81a4f8ba-127a-4d45-9fe2-462656e5ff9d", "address": "fa:16:3e:6d:14:43", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a4f8ba-12", "ovs_interfaceid": "81a4f8ba-127a-4d45-9fe2-462656e5ff9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.577781] env[62183]: DEBUG nova.compute.manager [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Received event network-vif-plugged-81a4f8ba-127a-4d45-9fe2-462656e5ff9d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.578154] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] Acquiring lock "143323db-da1c-4dd8-8f13-ed0af31e2027-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.578513] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] Lock "143323db-da1c-4dd8-8f13-ed0af31e2027-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.578835] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] Lock "143323db-da1c-4dd8-8f13-ed0af31e2027-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.579133] env[62183]: DEBUG nova.compute.manager [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] No waiting events found dispatching network-vif-plugged-81a4f8ba-127a-4d45-9fe2-462656e5ff9d {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 980.579421] env[62183]: WARNING nova.compute.manager [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Received unexpected event network-vif-plugged-81a4f8ba-127a-4d45-9fe2-462656e5ff9d for instance with vm_state building and task_state spawning. [ 980.579697] env[62183]: DEBUG nova.compute.manager [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Received event network-changed-81a4f8ba-127a-4d45-9fe2-462656e5ff9d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.579963] env[62183]: DEBUG nova.compute.manager [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Refreshing instance network info cache due to event network-changed-81a4f8ba-127a-4d45-9fe2-462656e5ff9d. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 980.580262] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] Acquiring lock "refresh_cache-143323db-da1c-4dd8-8f13-ed0af31e2027" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.625035] env[62183]: DEBUG nova.scheduler.client.report [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 980.639104] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.833020] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "refresh_cache-143323db-da1c-4dd8-8f13-ed0af31e2027" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.833366] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Instance network_info: |[{"id": "81a4f8ba-127a-4d45-9fe2-462656e5ff9d", "address": "fa:16:3e:6d:14:43", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a4f8ba-12", "ovs_interfaceid": "81a4f8ba-127a-4d45-9fe2-462656e5ff9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 980.833723] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] Acquired lock "refresh_cache-143323db-da1c-4dd8-8f13-ed0af31e2027" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.833913] env[62183]: DEBUG nova.network.neutron [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Refreshing network info cache for port 81a4f8ba-127a-4d45-9fe2-462656e5ff9d {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 980.835206] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:14:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4df917f7-847a-4c0e-b0e3-69a52e4a1554', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81a4f8ba-127a-4d45-9fe2-462656e5ff9d', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 980.842680] env[62183]: DEBUG oslo.service.loopingcall [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.843773] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 980.844026] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-368a24d4-4eb0-41fa-ada2-58a006cee6ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.866874] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 980.866874] env[62183]: value = "task-1387338" [ 980.866874] env[62183]: _type = "Task" [ 980.866874] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.877272] env[62183]: DEBUG nova.compute.manager [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 980.879015] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387338, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.903910] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.904206] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.904544] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.904778] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.904996] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.905200] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.905420] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.905586] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.905757] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.905922] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.906109] env[62183]: DEBUG nova.virt.hardware [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.907049] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8805df79-5805-48ef-8501-bfcf4ac32cf1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.910429] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.910596] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.910773] env[62183]: DEBUG nova.network.neutron [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 980.917304] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7509c428-8f37-4389-ad27-ba8b4efea62c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.133022] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.133022] env[62183]: DEBUG nova.compute.manager [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.135575] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.767s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.140152] env[62183]: INFO nova.compute.claims [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.145206] env[62183]: DEBUG nova.network.neutron [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Successfully updated port: f9702809-bb25-463e-a193-580acaeaf306 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 981.155394] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.378884] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387338, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.645405] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387337, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.647880] env[62183]: DEBUG nova.compute.utils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.658205] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquiring lock "refresh_cache-036037b3-9676-419c-ab87-e5a557438fbe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.658205] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquired lock "refresh_cache-036037b3-9676-419c-ab87-e5a557438fbe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.658205] env[62183]: DEBUG nova.network.neutron [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.658205] env[62183]: DEBUG nova.compute.manager [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.658205] env[62183]: DEBUG nova.network.neutron [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 981.704989] env[62183]: DEBUG nova.policy [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88efcb2057e84413806f9ed809c260e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eda7e297c551407295b3c67fa0f1dbc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.823120] env[62183]: DEBUG nova.network.neutron [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Updated VIF entry in instance network info cache for port 81a4f8ba-127a-4d45-9fe2-462656e5ff9d. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 981.823519] env[62183]: DEBUG nova.network.neutron [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Updating instance_info_cache with network_info: [{"id": "81a4f8ba-127a-4d45-9fe2-462656e5ff9d", "address": "fa:16:3e:6d:14:43", "network": {"id": "3c724a9e-d113-4ded-8f64-50351bd9a68f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1892268323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "499aba16038147e3b20b06876234cb13", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4df917f7-847a-4c0e-b0e3-69a52e4a1554", "external-id": "cl2-zone-457", "segmentation_id": 457, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a4f8ba-12", "ovs_interfaceid": "81a4f8ba-127a-4d45-9fe2-462656e5ff9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.872341] env[62183]: DEBUG nova.network.neutron [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [{"id": "3d687443-41fd-44ff-8981-35319b79d88a", "address": "fa:16:3e:91:31:9b", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d687443-41", "ovs_interfaceid": "3d687443-41fd-44ff-8981-35319b79d88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.879474] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387338, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.955539] env[62183]: DEBUG nova.network.neutron [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Successfully created port: 1fa5a463-ffcd-41f4-b61d-3fa88e6178c9 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 982.145474] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387337, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.530255} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.145964] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] b6402e2f-1035-4c1b-9a1e-05b17d89e4bb/b6402e2f-1035-4c1b-9a1e-05b17d89e4bb.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 982.145964] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.146236] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2bbf0e1b-4f04-4954-9652-67070262f178 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.153548] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 982.153548] env[62183]: value = "task-1387339" [ 982.153548] env[62183]: _type = "Task" [ 982.153548] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.156788] env[62183]: DEBUG nova.compute.manager [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 982.168427] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387339, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.199601] env[62183]: DEBUG nova.network.neutron [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 982.325908] env[62183]: DEBUG oslo_concurrency.lockutils [req-e7f70860-5cd1-4cf3-a282-ef8e774ea169 req-1f4a64ac-0258-4c22-a3cb-ab46a7649429 service nova] Releasing lock "refresh_cache-143323db-da1c-4dd8-8f13-ed0af31e2027" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.367929] env[62183]: DEBUG nova.network.neutron [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Updating instance_info_cache with network_info: [{"id": "f9702809-bb25-463e-a193-580acaeaf306", "address": "fa:16:3e:59:bb:43", "network": {"id": "b71bfd89-765a-466e-ba1d-7ac6e7f35ac2", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1468236707-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "218d7143fba34b8081e826b83968d53d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b8b5b5e2-866d-4ab5-b74d-4a47de0c4877", "external-id": "nsx-vlan-transportzone-74", "segmentation_id": 74, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9702809-bb", "ovs_interfaceid": "f9702809-bb25-463e-a193-580acaeaf306", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.374575] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.383646] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63548574-959c-4022-bc0b-82503c28866f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.386009] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387338, 'name': CreateVM_Task, 'duration_secs': 1.460232} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.386770] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 982.387730] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.387897] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.388222] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.388699] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04f16a1c-b30e-4530-a07f-1ea411d86098 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.392891] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a4d8f0-95d2-47a9-93c5-142afd8d1824 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.397275] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 982.397275] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5288047c-a1c4-93ec-4566-f6a979037e2c" [ 982.397275] env[62183]: _type = "Task" [ 982.397275] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.425890] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8a217f-51dc-4fe0-ad97-801e83f208a4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.431913] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5288047c-a1c4-93ec-4566-f6a979037e2c, 'name': SearchDatastore_Task, 'duration_secs': 0.010288} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.432560] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.432831] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.433100] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.433259] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.433443] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.433712] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebb428a9-0f87-4afb-9c16-248b7a8a6cb5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.438747] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392cf8ad-2ff8-44ba-860e-e750d2c0aaf3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.445150] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.445150] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 982.445150] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b154708a-d3c7-499a-b7a0-de9de1211924 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.455044] env[62183]: DEBUG nova.compute.provider_tree [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.459859] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 982.459859] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52ac9242-935f-7c12-f94d-bf9602fdc91e" [ 982.459859] env[62183]: _type = "Task" [ 982.459859] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.467644] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ac9242-935f-7c12-f94d-bf9602fdc91e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.602843] env[62183]: DEBUG nova.compute.manager [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Received event network-vif-plugged-f9702809-bb25-463e-a193-580acaeaf306 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.603121] env[62183]: DEBUG oslo_concurrency.lockutils [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] Acquiring lock "036037b3-9676-419c-ab87-e5a557438fbe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.603344] env[62183]: DEBUG oslo_concurrency.lockutils [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] Lock "036037b3-9676-419c-ab87-e5a557438fbe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.603457] env[62183]: DEBUG oslo_concurrency.lockutils [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] Lock "036037b3-9676-419c-ab87-e5a557438fbe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.603620] env[62183]: DEBUG nova.compute.manager [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] No waiting events found dispatching network-vif-plugged-f9702809-bb25-463e-a193-580acaeaf306 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 982.603764] env[62183]: WARNING nova.compute.manager [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Received unexpected event network-vif-plugged-f9702809-bb25-463e-a193-580acaeaf306 for instance with vm_state building and task_state spawning. [ 982.603923] env[62183]: DEBUG nova.compute.manager [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Received event network-changed-f9702809-bb25-463e-a193-580acaeaf306 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.604311] env[62183]: DEBUG nova.compute.manager [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Refreshing instance network info cache due to event network-changed-f9702809-bb25-463e-a193-580acaeaf306. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.604519] env[62183]: DEBUG oslo_concurrency.lockutils [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] Acquiring lock "refresh_cache-036037b3-9676-419c-ab87-e5a557438fbe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.667219] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387339, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079688} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.667769] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.668552] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505bf536-e7d3-4bd8-aa14-51e0b1de36ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.690818] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] b6402e2f-1035-4c1b-9a1e-05b17d89e4bb/b6402e2f-1035-4c1b-9a1e-05b17d89e4bb.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.691135] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72b6af57-f1f4-43a5-91ca-de77841ca205 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.713052] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 982.713052] env[62183]: value = "task-1387340" [ 982.713052] env[62183]: _type = "Task" [ 982.713052] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.721606] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387340, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.873767] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Releasing lock "refresh_cache-036037b3-9676-419c-ab87-e5a557438fbe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.874118] env[62183]: DEBUG nova.compute.manager [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Instance network_info: |[{"id": "f9702809-bb25-463e-a193-580acaeaf306", "address": "fa:16:3e:59:bb:43", "network": {"id": "b71bfd89-765a-466e-ba1d-7ac6e7f35ac2", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1468236707-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "218d7143fba34b8081e826b83968d53d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b8b5b5e2-866d-4ab5-b74d-4a47de0c4877", "external-id": "nsx-vlan-transportzone-74", "segmentation_id": 74, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9702809-bb", "ovs_interfaceid": "f9702809-bb25-463e-a193-580acaeaf306", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 982.874484] env[62183]: DEBUG oslo_concurrency.lockutils [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] Acquired lock "refresh_cache-036037b3-9676-419c-ab87-e5a557438fbe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.874679] env[62183]: DEBUG nova.network.neutron [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Refreshing network info cache for port f9702809-bb25-463e-a193-580acaeaf306 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.875919] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:bb:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b8b5b5e2-866d-4ab5-b74d-4a47de0c4877', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9702809-bb25-463e-a193-580acaeaf306', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.885013] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Creating folder: Project (218d7143fba34b8081e826b83968d53d). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 982.893356] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cefc0df7-d1dc-4b11-96bc-1fa70a881c63 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.905139] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Created folder: Project (218d7143fba34b8081e826b83968d53d) in parent group-v294392. [ 982.905396] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Creating folder: Instances. Parent ref: group-v294502. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 982.905720] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c81e34d3-8fac-4e9d-8db5-4bd7e209f651 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.908974] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999daec6-e5ad-4403-9025-689e87a5966e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.930226] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c886404-28b4-437b-b3d1-90cf3ec39118 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.932897] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Created folder: Instances in parent group-v294502. [ 982.933192] env[62183]: DEBUG oslo.service.loopingcall [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.933417] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.934015] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1272a534-d94d-4a4c-864d-47cb7925be2d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.952203] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance 'a948464b-63aa-4bc8-9885-228049e96d37' progress to 83 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 982.958274] env[62183]: DEBUG nova.scheduler.client.report [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 982.962551] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.962551] env[62183]: value = "task-1387343" [ 982.962551] env[62183]: _type = "Task" [ 982.962551] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.975365] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ac9242-935f-7c12-f94d-bf9602fdc91e, 'name': SearchDatastore_Task, 'duration_secs': 0.009892} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.979020] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387343, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.979852] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fd764d1-6029-4956-8577-f70fdf11a5df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.986773] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 982.986773] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52ef45f6-2f02-86d9-667b-92ad36d0bcf3" [ 982.986773] env[62183]: _type = "Task" [ 982.986773] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.995446] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ef45f6-2f02-86d9-667b-92ad36d0bcf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.171396] env[62183]: DEBUG nova.compute.manager [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.197882] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.198142] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.198304] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.198493] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.198644] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.198789] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.198996] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.199173] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.199343] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.199508] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.199686] env[62183]: DEBUG nova.virt.hardware [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.200567] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f72a7d-0dea-4ffa-b793-e291d006c722 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.208828] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c883429-06eb-4082-baa5-bd8cf6a2f333 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.230131] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.352469] env[62183]: DEBUG nova.compute.manager [req-17068d66-5fdc-40fc-b350-2d32962b572a req-44ab5726-c8ec-471a-b0ca-4cafe20d96fd service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Received event network-vif-plugged-1fa5a463-ffcd-41f4-b61d-3fa88e6178c9 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.352708] env[62183]: DEBUG oslo_concurrency.lockutils [req-17068d66-5fdc-40fc-b350-2d32962b572a req-44ab5726-c8ec-471a-b0ca-4cafe20d96fd service nova] Acquiring lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.352919] env[62183]: DEBUG oslo_concurrency.lockutils [req-17068d66-5fdc-40fc-b350-2d32962b572a req-44ab5726-c8ec-471a-b0ca-4cafe20d96fd service nova] Lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.353331] env[62183]: DEBUG oslo_concurrency.lockutils [req-17068d66-5fdc-40fc-b350-2d32962b572a req-44ab5726-c8ec-471a-b0ca-4cafe20d96fd service nova] Lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.353542] env[62183]: DEBUG nova.compute.manager [req-17068d66-5fdc-40fc-b350-2d32962b572a req-44ab5726-c8ec-471a-b0ca-4cafe20d96fd service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] No waiting events found dispatching network-vif-plugged-1fa5a463-ffcd-41f4-b61d-3fa88e6178c9 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 983.353719] env[62183]: WARNING nova.compute.manager [req-17068d66-5fdc-40fc-b350-2d32962b572a req-44ab5726-c8ec-471a-b0ca-4cafe20d96fd service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Received unexpected event network-vif-plugged-1fa5a463-ffcd-41f4-b61d-3fa88e6178c9 for instance with vm_state building and task_state spawning. [ 983.460083] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b924f0-088d-4792-93d3-f5ee18d6d757 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance 'a948464b-63aa-4bc8-9885-228049e96d37' progress to 100 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 983.465669] env[62183]: DEBUG nova.network.neutron [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Successfully updated port: 1fa5a463-ffcd-41f4-b61d-3fa88e6178c9 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.467197] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.467701] env[62183]: DEBUG nova.compute.manager [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.479629] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.584s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.479629] env[62183]: DEBUG nova.objects.instance [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lazy-loading 'resources' on Instance uuid 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.480986] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387343, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.497869] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ef45f6-2f02-86d9-667b-92ad36d0bcf3, 'name': SearchDatastore_Task, 'duration_secs': 0.010368} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.498265] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.498536] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 143323db-da1c-4dd8-8f13-ed0af31e2027/143323db-da1c-4dd8-8f13-ed0af31e2027.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 983.498802] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0144d64d-dd0e-4e50-aa51-f9d6948eabec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.507061] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 983.507061] env[62183]: value = "task-1387344" [ 983.507061] env[62183]: _type = "Task" [ 983.507061] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.515494] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387344, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.616182] env[62183]: DEBUG nova.network.neutron [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Updated VIF entry in instance network info cache for port f9702809-bb25-463e-a193-580acaeaf306. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.616626] env[62183]: DEBUG nova.network.neutron [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Updating instance_info_cache with network_info: [{"id": "f9702809-bb25-463e-a193-580acaeaf306", "address": "fa:16:3e:59:bb:43", "network": {"id": "b71bfd89-765a-466e-ba1d-7ac6e7f35ac2", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1468236707-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "218d7143fba34b8081e826b83968d53d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b8b5b5e2-866d-4ab5-b74d-4a47de0c4877", "external-id": "nsx-vlan-transportzone-74", "segmentation_id": 74, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9702809-bb", "ovs_interfaceid": "f9702809-bb25-463e-a193-580acaeaf306", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.724234] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387340, 'name': ReconfigVM_Task, 'duration_secs': 0.639968} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.724654] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Reconfigured VM instance instance-0000005d to attach disk [datastore1] b6402e2f-1035-4c1b-9a1e-05b17d89e4bb/b6402e2f-1035-4c1b-9a1e-05b17d89e4bb.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.725202] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3794cb22-52ec-4052-bc0a-906b2781dc45 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.733729] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 983.733729] env[62183]: value = "task-1387345" [ 983.733729] env[62183]: _type = "Task" [ 983.733729] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.743559] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387345, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.972664] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "refresh_cache-24f7c072-1209-4c6e-9a54-4e069f14f7d2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.972664] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "refresh_cache-24f7c072-1209-4c6e-9a54-4e069f14f7d2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.972664] env[62183]: DEBUG nova.network.neutron [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 983.976909] env[62183]: DEBUG nova.compute.utils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.985842] env[62183]: DEBUG nova.compute.manager [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.986050] env[62183]: DEBUG nova.network.neutron [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.987945] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387343, 'name': CreateVM_Task, 'duration_secs': 0.545503} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.988727] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 983.989660] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.989834] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.990443] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 983.991054] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c082b2df-fd67-406c-9e49-b5242c15974c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.998504] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 983.998504] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5288c206-47da-c27e-4af2-5845cc9c2ae1" [ 983.998504] env[62183]: _type = "Task" [ 983.998504] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.015051] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5288c206-47da-c27e-4af2-5845cc9c2ae1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.021658] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387344, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.032459] env[62183]: DEBUG nova.policy [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b85b5dce9bf44da6afd11f100a121f03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '34581a9b5c1943eabc13bbb300a0f086', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.118817] env[62183]: DEBUG oslo_concurrency.lockutils [req-93ecf4fc-ea2f-4230-9887-54cf83b03301 req-31edf21c-2a51-4f8e-b547-b1daaba579e6 service nova] Releasing lock "refresh_cache-036037b3-9676-419c-ab87-e5a557438fbe" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.178852] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13223258-3734-49eb-aef8-e100eeed5903 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.187257] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061631c7-fc57-4252-ad64-6c8c1d699a7d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.217062] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252c6e48-27cd-4fda-ad9d-5853a7a67009 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.224815] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142b7831-5c70-4056-99c7-1a9f37d7269a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.238086] env[62183]: DEBUG nova.compute.provider_tree [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.247232] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387345, 'name': Rename_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.281592] env[62183]: DEBUG nova.network.neutron [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Successfully created port: 8b2e8106-ed68-4023-b3cc-d91842d269be {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.486383] env[62183]: DEBUG nova.compute.manager [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.521721] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5288c206-47da-c27e-4af2-5845cc9c2ae1, 'name': SearchDatastore_Task, 'duration_secs': 0.018235} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.523214] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.523579] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 984.523731] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.523887] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.524084] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 984.524629] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aca23662-5c0b-40d5-91d1-6277630dac1c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.530496] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387344, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.531570] env[62183]: DEBUG nova.network.neutron [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 984.541091] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 984.541304] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 984.542043] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f258a94-0530-45c1-979c-0453fc1330b3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.551967] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 984.551967] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e9535c-bc8c-bdcf-9214-6b09405f3afc" [ 984.551967] env[62183]: _type = "Task" [ 984.551967] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.565217] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e9535c-bc8c-bdcf-9214-6b09405f3afc, 'name': SearchDatastore_Task, 'duration_secs': 0.010625} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.566198] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4396befe-3919-4611-8600-95304dc3ecc1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.573014] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 984.573014] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52b60c2a-5847-c0c1-a796-4151d3ca2dd7" [ 984.573014] env[62183]: _type = "Task" [ 984.573014] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.582670] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b60c2a-5847-c0c1-a796-4151d3ca2dd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.743261] env[62183]: DEBUG nova.scheduler.client.report [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 984.749856] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387345, 'name': Rename_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.893420] env[62183]: DEBUG nova.network.neutron [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Updating instance_info_cache with network_info: [{"id": "1fa5a463-ffcd-41f4-b61d-3fa88e6178c9", "address": "fa:16:3e:06:ae:b7", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fa5a463-ff", "ovs_interfaceid": "1fa5a463-ffcd-41f4-b61d-3fa88e6178c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.019132] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387344, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.083380] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b60c2a-5847-c0c1-a796-4151d3ca2dd7, 'name': SearchDatastore_Task, 'duration_secs': 0.010402} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.083640] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.083904] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 036037b3-9676-419c-ab87-e5a557438fbe/036037b3-9676-419c-ab87-e5a557438fbe.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 985.084212] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1ec1cbb-7e1b-4661-a574-a414e8b50660 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.093058] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 985.093058] env[62183]: value = "task-1387346" [ 985.093058] env[62183]: _type = "Task" [ 985.093058] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.101319] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387346, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.248690] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387345, 'name': Rename_Task, 'duration_secs': 1.192188} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.248969] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 985.249226] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-953f5bf1-daac-414c-ba6d-dc2932fe8857 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.251262] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.778s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.253330] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.062s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.253518] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.253696] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 985.254778] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f7812e-473f-4c59-8e99-d98b666b209c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.260233] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 985.260233] env[62183]: value = "task-1387347" [ 985.260233] env[62183]: _type = "Task" [ 985.260233] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.266582] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b00e4e6-2621-46c2-a74b-84f053e6c778 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.284764] env[62183]: INFO nova.scheduler.client.report [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Deleted allocations for instance 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec [ 985.286800] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636e3c82-071a-4da3-97d1-06d61adda7f9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.289609] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.297756] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47eeda25-a1d0-4c95-b8a8-31f2e5fd737a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.335536] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180332MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 985.335732] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.335897] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.380030] env[62183]: DEBUG nova.compute.manager [req-d1a30cee-985f-42f0-b2ff-d0406d66ea4f req-89280e7a-b717-4422-b55d-0bf148c238c7 service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Received event network-changed-1fa5a463-ffcd-41f4-b61d-3fa88e6178c9 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.380030] env[62183]: DEBUG nova.compute.manager [req-d1a30cee-985f-42f0-b2ff-d0406d66ea4f req-89280e7a-b717-4422-b55d-0bf148c238c7 service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Refreshing instance network info cache due to event network-changed-1fa5a463-ffcd-41f4-b61d-3fa88e6178c9. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 985.380030] env[62183]: DEBUG oslo_concurrency.lockutils [req-d1a30cee-985f-42f0-b2ff-d0406d66ea4f req-89280e7a-b717-4422-b55d-0bf148c238c7 service nova] Acquiring lock "refresh_cache-24f7c072-1209-4c6e-9a54-4e069f14f7d2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.397765] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "refresh_cache-24f7c072-1209-4c6e-9a54-4e069f14f7d2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.398127] env[62183]: DEBUG nova.compute.manager [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Instance network_info: |[{"id": "1fa5a463-ffcd-41f4-b61d-3fa88e6178c9", "address": "fa:16:3e:06:ae:b7", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fa5a463-ff", "ovs_interfaceid": "1fa5a463-ffcd-41f4-b61d-3fa88e6178c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 985.398458] env[62183]: DEBUG oslo_concurrency.lockutils [req-d1a30cee-985f-42f0-b2ff-d0406d66ea4f req-89280e7a-b717-4422-b55d-0bf148c238c7 service nova] Acquired lock "refresh_cache-24f7c072-1209-4c6e-9a54-4e069f14f7d2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.398708] env[62183]: DEBUG nova.network.neutron [req-d1a30cee-985f-42f0-b2ff-d0406d66ea4f req-89280e7a-b717-4422-b55d-0bf148c238c7 service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Refreshing network info cache for port 1fa5a463-ffcd-41f4-b61d-3fa88e6178c9 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.400239] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:ae:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1fa5a463-ffcd-41f4-b61d-3fa88e6178c9', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.408710] env[62183]: DEBUG oslo.service.loopingcall [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.409719] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.409985] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69cfa220-975d-432f-aea2-0454ab9c2b0f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.433423] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.433423] env[62183]: value = "task-1387348" [ 985.433423] env[62183]: _type = "Task" [ 985.433423] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.442939] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387348, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.494866] env[62183]: DEBUG nova.compute.manager [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.521378] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387344, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.522507} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.524512] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 143323db-da1c-4dd8-8f13-ed0af31e2027/143323db-da1c-4dd8-8f13-ed0af31e2027.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 985.524751] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.525614] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1f7de04-c415-4b40-be7d-d2bd492d01df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.532886] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.533244] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.533438] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.533630] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.533779] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.533956] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.534173] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.534350] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.534548] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.534735] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.534891] env[62183]: DEBUG nova.virt.hardware [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.535763] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a259e3-8dee-4dba-8595-23dc159071be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.539988] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 985.539988] env[62183]: value = "task-1387349" [ 985.539988] env[62183]: _type = "Task" [ 985.539988] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.547853] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899370da-5282-4faa-b623-b1925a734f78 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.554767] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387349, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.604172] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387346, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.774819] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387347, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.800023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-8db384f2-fa1e-4b02-b3dc-d8e9488e36f0 tempest-ServerDiskConfigTestJSON-4200209 tempest-ServerDiskConfigTestJSON-4200209-project-member] Lock "8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.835s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.877546] env[62183]: DEBUG nova.compute.manager [req-3893e0e3-f6eb-47e1-acc8-0899866180a6 req-04ba7b39-4cbf-4ac7-a827-574aa4b9c376 service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Received event network-vif-plugged-8b2e8106-ed68-4023-b3cc-d91842d269be {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.877546] env[62183]: DEBUG oslo_concurrency.lockutils [req-3893e0e3-f6eb-47e1-acc8-0899866180a6 req-04ba7b39-4cbf-4ac7-a827-574aa4b9c376 service nova] Acquiring lock "d8817d24-b0cb-4956-b195-cc417ae09fb4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.877546] env[62183]: DEBUG oslo_concurrency.lockutils [req-3893e0e3-f6eb-47e1-acc8-0899866180a6 req-04ba7b39-4cbf-4ac7-a827-574aa4b9c376 service nova] Lock "d8817d24-b0cb-4956-b195-cc417ae09fb4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.877732] env[62183]: DEBUG oslo_concurrency.lockutils [req-3893e0e3-f6eb-47e1-acc8-0899866180a6 req-04ba7b39-4cbf-4ac7-a827-574aa4b9c376 service nova] Lock "d8817d24-b0cb-4956-b195-cc417ae09fb4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.877732] env[62183]: DEBUG nova.compute.manager [req-3893e0e3-f6eb-47e1-acc8-0899866180a6 req-04ba7b39-4cbf-4ac7-a827-574aa4b9c376 service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] No waiting events found dispatching network-vif-plugged-8b2e8106-ed68-4023-b3cc-d91842d269be {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 985.877895] env[62183]: WARNING nova.compute.manager [req-3893e0e3-f6eb-47e1-acc8-0899866180a6 req-04ba7b39-4cbf-4ac7-a827-574aa4b9c376 service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Received unexpected event network-vif-plugged-8b2e8106-ed68-4023-b3cc-d91842d269be for instance with vm_state building and task_state spawning. [ 985.944607] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387348, 'name': CreateVM_Task, 'duration_secs': 0.370807} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.945128] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.945810] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.945985] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.946323] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.946611] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f104fa62-cbd3-4a8b-b115-aebb389ed43c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.951951] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 985.951951] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52ea8b1a-6c46-91c8-2103-f85a6fa0df76" [ 985.951951] env[62183]: _type = "Task" [ 985.951951] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.961378] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ea8b1a-6c46-91c8-2103-f85a6fa0df76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.990308] env[62183]: DEBUG nova.network.neutron [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Successfully updated port: 8b2e8106-ed68-4023-b3cc-d91842d269be {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.051656] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387349, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086805} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.051656] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.052608] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915a891b-dae8-446e-8bd7-8f40f533b3c7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.076812] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 143323db-da1c-4dd8-8f13-ed0af31e2027/143323db-da1c-4dd8-8f13-ed0af31e2027.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.080076] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c95ecee3-e0d3-41a2-bbf6-7c75c719bbbd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.104264] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387346, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.70176} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.105045] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 036037b3-9676-419c-ab87-e5a557438fbe/036037b3-9676-419c-ab87-e5a557438fbe.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 986.105293] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 986.106216] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6dcce3b7-1684-4fd2-92ac-b894cc5aa98f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.108231] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 986.108231] env[62183]: value = "task-1387350" [ 986.108231] env[62183]: _type = "Task" [ 986.108231] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.113262] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 986.113262] env[62183]: value = "task-1387351" [ 986.113262] env[62183]: _type = "Task" [ 986.113262] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.119754] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387350, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.126488] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387351, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.190344] env[62183]: DEBUG nova.network.neutron [req-d1a30cee-985f-42f0-b2ff-d0406d66ea4f req-89280e7a-b717-4422-b55d-0bf148c238c7 service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Updated VIF entry in instance network info cache for port 1fa5a463-ffcd-41f4-b61d-3fa88e6178c9. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.190765] env[62183]: DEBUG nova.network.neutron [req-d1a30cee-985f-42f0-b2ff-d0406d66ea4f req-89280e7a-b717-4422-b55d-0bf148c238c7 service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Updating instance_info_cache with network_info: [{"id": "1fa5a463-ffcd-41f4-b61d-3fa88e6178c9", "address": "fa:16:3e:06:ae:b7", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fa5a463-ff", "ovs_interfaceid": "1fa5a463-ffcd-41f4-b61d-3fa88e6178c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.274571] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387347, 'name': PowerOnVM_Task, 'duration_secs': 0.559084} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.275180] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 986.275180] env[62183]: INFO nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Took 9.96 seconds to spawn the instance on the hypervisor. [ 986.275551] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 986.276430] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81346d96-f3cc-4d3b-a141-7078282c917d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.348456] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Applying migration context for instance a948464b-63aa-4bc8-9885-228049e96d37 as it has an incoming, in-progress migration 4e58e303-0d2b-4df3-96ab-52a8e649d12b. Migration status is confirming {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 986.348759] env[62183]: INFO nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating resource usage from migration 4e58e303-0d2b-4df3-96ab-52a8e649d12b [ 986.360936] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "a948464b-63aa-4bc8-9885-228049e96d37" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.361237] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.361470] env[62183]: DEBUG nova.compute.manager [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Going to confirm migration 4 {{(pid=62183) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 986.372301] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.372450] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.372574] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.372729] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance fb348784-62a0-4d1f-ac7f-f176f3da0dd9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.372897] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.373108] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 8c4b68d6-fedd-408f-a449-aace7400014a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.373353] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 3a3ca703-2883-4aa9-a33f-326cc25d4838 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.373531] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Migration 4e58e303-0d2b-4df3-96ab-52a8e649d12b is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 986.373805] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a948464b-63aa-4bc8-9885-228049e96d37 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.373933] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance b6402e2f-1035-4c1b-9a1e-05b17d89e4bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.374096] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 143323db-da1c-4dd8-8f13-ed0af31e2027 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.374234] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 036037b3-9676-419c-ab87-e5a557438fbe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.374350] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 24f7c072-1209-4c6e-9a54-4e069f14f7d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.374465] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance d8817d24-b0cb-4956-b195-cc417ae09fb4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.374666] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 986.374814] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 986.463113] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ea8b1a-6c46-91c8-2103-f85a6fa0df76, 'name': SearchDatastore_Task, 'duration_secs': 0.010117} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.465462] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.465707] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.465940] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.466104] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.466283] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.467374] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71b383b8-3b57-442a-896b-f84b7b36e4ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.475725] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.475909] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 986.478919] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79efa28c-4ade-4f0a-b20e-66bde0d4b206 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.484702] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 986.484702] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5252cacc-4736-91bf-a132-12d83ba91eae" [ 986.484702] env[62183]: _type = "Task" [ 986.484702] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.492640] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-d8817d24-b0cb-4956-b195-cc417ae09fb4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.492772] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-d8817d24-b0cb-4956-b195-cc417ae09fb4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.492897] env[62183]: DEBUG nova.network.neutron [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.493968] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5252cacc-4736-91bf-a132-12d83ba91eae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.573444] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f50678-55cc-4673-91e8-00e13d2b62b3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.581396] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5034bcc3-da18-4269-8f78-a0069c369366 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.613987] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94483f6a-d92f-41d9-9e56-767c582a9c1d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.623083] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387350, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.626403] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a922d67-2fe1-4563-82c7-07671c0b5435 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.632535] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387351, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073439} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.633087] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.633840] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9be520e-b02c-4387-9f71-62c9961b2594 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.645061] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.665491] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 036037b3-9676-419c-ab87-e5a557438fbe/036037b3-9676-419c-ab87-e5a557438fbe.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.666341] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9d160ed-aebe-4461-9b89-52c7846d841a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.686333] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 986.686333] env[62183]: value = "task-1387352" [ 986.686333] env[62183]: _type = "Task" [ 986.686333] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.696145] env[62183]: DEBUG oslo_concurrency.lockutils [req-d1a30cee-985f-42f0-b2ff-d0406d66ea4f req-89280e7a-b717-4422-b55d-0bf148c238c7 service nova] Releasing lock "refresh_cache-24f7c072-1209-4c6e-9a54-4e069f14f7d2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.696593] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387352, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.797018] env[62183]: INFO nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Took 14.80 seconds to build instance. [ 986.902308] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.902523] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.902723] env[62183]: DEBUG nova.network.neutron [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.902909] env[62183]: DEBUG nova.objects.instance [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'info_cache' on Instance uuid a948464b-63aa-4bc8-9885-228049e96d37 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.995983] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5252cacc-4736-91bf-a132-12d83ba91eae, 'name': SearchDatastore_Task, 'duration_secs': 0.008666} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.998490] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d97aa283-14f7-49fc-92d8-4726ed462ab9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.004214] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 987.004214] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]527be5e0-4159-c433-ab29-6d8066bac42f" [ 987.004214] env[62183]: _type = "Task" [ 987.004214] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.012313] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527be5e0-4159-c433-ab29-6d8066bac42f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.031170] env[62183]: DEBUG nova.network.neutron [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 987.127539] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387350, 'name': ReconfigVM_Task, 'duration_secs': 0.879239} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.129030] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 143323db-da1c-4dd8-8f13-ed0af31e2027/143323db-da1c-4dd8-8f13-ed0af31e2027.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.129722] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e7525aa-e6fd-4f92-9c7f-0df4c1911aab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.139311] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 987.139311] env[62183]: value = "task-1387353" [ 987.139311] env[62183]: _type = "Task" [ 987.139311] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.148693] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 987.151890] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387353, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.197879] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387352, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.212605] env[62183]: DEBUG nova.network.neutron [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Updating instance_info_cache with network_info: [{"id": "8b2e8106-ed68-4023-b3cc-d91842d269be", "address": "fa:16:3e:ea:dd:c5", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2e8106-ed", "ovs_interfaceid": "8b2e8106-ed68-4023-b3cc-d91842d269be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.299216] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.313s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.517647] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527be5e0-4159-c433-ab29-6d8066bac42f, 'name': SearchDatastore_Task, 'duration_secs': 0.010494} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.517647] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.517647] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 24f7c072-1209-4c6e-9a54-4e069f14f7d2/24f7c072-1209-4c6e-9a54-4e069f14f7d2.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.517868] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd25fd9c-95b6-4938-93f8-400244fa16cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.525870] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 987.525870] env[62183]: value = "task-1387354" [ 987.525870] env[62183]: _type = "Task" [ 987.525870] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.537762] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387354, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.649615] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387353, 'name': Rename_Task, 'duration_secs': 0.160897} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.650028] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 987.650323] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb31fccc-cc4c-4dd4-bbff-aa3049d06915 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.652906] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 987.653143] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.317s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.656992] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 987.656992] env[62183]: value = "task-1387355" [ 987.656992] env[62183]: _type = "Task" [ 987.656992] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.664879] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387355, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.697844] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387352, 'name': ReconfigVM_Task, 'duration_secs': 0.652991} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.697984] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 036037b3-9676-419c-ab87-e5a557438fbe/036037b3-9676-419c-ab87-e5a557438fbe.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 987.698619] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c35d9b9-8052-49ae-8048-d2fb6b945ca5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.707501] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 987.707501] env[62183]: value = "task-1387356" [ 987.707501] env[62183]: _type = "Task" [ 987.707501] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.718196] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-d8817d24-b0cb-4956-b195-cc417ae09fb4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.718390] env[62183]: DEBUG nova.compute.manager [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Instance network_info: |[{"id": "8b2e8106-ed68-4023-b3cc-d91842d269be", "address": "fa:16:3e:ea:dd:c5", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2e8106-ed", "ovs_interfaceid": "8b2e8106-ed68-4023-b3cc-d91842d269be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.718706] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387356, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.719137] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:dd:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8800a981-a89e-42e4-8be9-cace419ba9cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b2e8106-ed68-4023-b3cc-d91842d269be', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.726715] env[62183]: DEBUG oslo.service.loopingcall [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.727014] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.727254] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63308148-0a97-4869-93d9-6a08261d6db3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.749986] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.749986] env[62183]: value = "task-1387357" [ 987.749986] env[62183]: _type = "Task" [ 987.749986] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.761442] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387357, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.914948] env[62183]: DEBUG nova.compute.manager [req-b50d8ad4-e243-4905-81a1-1c0948526a6c req-a7ca8544-b503-4748-8703-0e598c5e55fa service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Received event network-changed-8b2e8106-ed68-4023-b3cc-d91842d269be {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.915325] env[62183]: DEBUG nova.compute.manager [req-b50d8ad4-e243-4905-81a1-1c0948526a6c req-a7ca8544-b503-4748-8703-0e598c5e55fa service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Refreshing instance network info cache due to event network-changed-8b2e8106-ed68-4023-b3cc-d91842d269be. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 987.915711] env[62183]: DEBUG oslo_concurrency.lockutils [req-b50d8ad4-e243-4905-81a1-1c0948526a6c req-a7ca8544-b503-4748-8703-0e598c5e55fa service nova] Acquiring lock "refresh_cache-d8817d24-b0cb-4956-b195-cc417ae09fb4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.915947] env[62183]: DEBUG oslo_concurrency.lockutils [req-b50d8ad4-e243-4905-81a1-1c0948526a6c req-a7ca8544-b503-4748-8703-0e598c5e55fa service nova] Acquired lock "refresh_cache-d8817d24-b0cb-4956-b195-cc417ae09fb4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.916214] env[62183]: DEBUG nova.network.neutron [req-b50d8ad4-e243-4905-81a1-1c0948526a6c req-a7ca8544-b503-4748-8703-0e598c5e55fa service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Refreshing network info cache for port 8b2e8106-ed68-4023-b3cc-d91842d269be {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.035871] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387354, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.148168] env[62183]: DEBUG nova.network.neutron [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [{"id": "3d687443-41fd-44ff-8981-35319b79d88a", "address": "fa:16:3e:91:31:9b", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d687443-41", "ovs_interfaceid": "3d687443-41fd-44ff-8981-35319b79d88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.169844] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387355, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.222392] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387356, 'name': Rename_Task, 'duration_secs': 0.176091} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.222845] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 988.222965] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8281f613-57c8-4a66-9667-7cd14dfda14b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.229967] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 988.229967] env[62183]: value = "task-1387358" [ 988.229967] env[62183]: _type = "Task" [ 988.229967] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.245406] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387358, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.259597] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387357, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.539381] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387354, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516829} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.539689] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 24f7c072-1209-4c6e-9a54-4e069f14f7d2/24f7c072-1209-4c6e-9a54-4e069f14f7d2.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.540056] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.540353] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f5c5e003-8f18-48ed-bae7-a54d7ac5a825 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.547431] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 988.547431] env[62183]: value = "task-1387359" [ 988.547431] env[62183]: _type = "Task" [ 988.547431] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.560671] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387359, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.651133] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.651456] env[62183]: DEBUG nova.objects.instance [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'migration_context' on Instance uuid a948464b-63aa-4bc8-9885-228049e96d37 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.669176] env[62183]: DEBUG oslo_vmware.api [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387355, 'name': PowerOnVM_Task, 'duration_secs': 0.530826} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.669583] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 988.669914] env[62183]: INFO nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Took 10.07 seconds to spawn the instance on the hypervisor. [ 988.670197] env[62183]: DEBUG nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.671215] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fcd9ab-81eb-48a4-b6f7-eb5c2a14d677 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.695701] env[62183]: DEBUG nova.network.neutron [req-b50d8ad4-e243-4905-81a1-1c0948526a6c req-a7ca8544-b503-4748-8703-0e598c5e55fa service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Updated VIF entry in instance network info cache for port 8b2e8106-ed68-4023-b3cc-d91842d269be. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 988.696046] env[62183]: DEBUG nova.network.neutron [req-b50d8ad4-e243-4905-81a1-1c0948526a6c req-a7ca8544-b503-4748-8703-0e598c5e55fa service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Updating instance_info_cache with network_info: [{"id": "8b2e8106-ed68-4023-b3cc-d91842d269be", "address": "fa:16:3e:ea:dd:c5", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b2e8106-ed", "ovs_interfaceid": "8b2e8106-ed68-4023-b3cc-d91842d269be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.740924] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387358, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.760576] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387357, 'name': CreateVM_Task, 'duration_secs': 0.580799} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.760762] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.761543] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.761597] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.762029] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.762171] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04ec3082-8f0c-40ea-ab14-ae543c569fde {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.767665] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 988.767665] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52526c34-794d-a66b-315f-c0eaa1486127" [ 988.767665] env[62183]: _type = "Task" [ 988.767665] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.776866] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52526c34-794d-a66b-315f-c0eaa1486127, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.060100] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387359, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102382} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.060450] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.061253] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0c6242-c8e3-45f8-a027-b0028f4b8001 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.084030] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 24f7c072-1209-4c6e-9a54-4e069f14f7d2/24f7c072-1209-4c6e-9a54-4e069f14f7d2.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.084369] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bafde315-1e94-4b41-b05c-5bb28871dc8d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.105013] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 989.105013] env[62183]: value = "task-1387360" [ 989.105013] env[62183]: _type = "Task" [ 989.105013] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.113474] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387360, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.159739] env[62183]: DEBUG nova.objects.base [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 989.160805] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed83564-59cf-44e8-8509-d6319f812d89 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.181294] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-015723ae-5246-41d4-8b4b-7b5a22266830 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.192543] env[62183]: INFO nova.compute.manager [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Took 17.17 seconds to build instance. [ 989.194105] env[62183]: DEBUG oslo_vmware.api [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 989.194105] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]522d2de6-6a86-7e17-d1d7-5e9647b1efd8" [ 989.194105] env[62183]: _type = "Task" [ 989.194105] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.198098] env[62183]: DEBUG oslo_concurrency.lockutils [req-b50d8ad4-e243-4905-81a1-1c0948526a6c req-a7ca8544-b503-4748-8703-0e598c5e55fa service nova] Releasing lock "refresh_cache-d8817d24-b0cb-4956-b195-cc417ae09fb4" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.205489] env[62183]: DEBUG oslo_vmware.api [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]522d2de6-6a86-7e17-d1d7-5e9647b1efd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.242129] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387358, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.282854] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52526c34-794d-a66b-315f-c0eaa1486127, 'name': SearchDatastore_Task, 'duration_secs': 0.009603} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.283202] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.283446] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.283838] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.283838] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.284077] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.284374] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75e84ee6-33c1-4124-8321-f7e08c61848e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.295272] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.295413] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.296322] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56bf6854-06c0-4d94-95c7-a6213a71bce0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.302830] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 989.302830] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f4b750-2287-08e5-ab8e-ee9079402a3c" [ 989.302830] env[62183]: _type = "Task" [ 989.302830] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.312542] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f4b750-2287-08e5-ab8e-ee9079402a3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.616279] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387360, 'name': ReconfigVM_Task, 'duration_secs': 0.380607} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.616612] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 24f7c072-1209-4c6e-9a54-4e069f14f7d2/24f7c072-1209-4c6e-9a54-4e069f14f7d2.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.617377] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-60928d4b-e088-4cce-9c74-14733490e241 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.624892] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 989.624892] env[62183]: value = "task-1387361" [ 989.624892] env[62183]: _type = "Task" [ 989.624892] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.639019] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387361, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.697031] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7bf8ddb5-9dfc-44a8-b0b0-9aac0b8a0c1e tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "143323db-da1c-4dd8-8f13-ed0af31e2027" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.679s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.710779] env[62183]: DEBUG oslo_vmware.api [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]522d2de6-6a86-7e17-d1d7-5e9647b1efd8, 'name': SearchDatastore_Task, 'duration_secs': 0.013142} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.711138] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.711423] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.728177] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "a5018b4e-016d-4c61-bfbf-2c039ab79499" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.728721] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.746678] env[62183]: DEBUG oslo_vmware.api [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387358, 'name': PowerOnVM_Task, 'duration_secs': 1.03145} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.747130] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 989.747623] env[62183]: INFO nova.compute.manager [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Took 8.87 seconds to spawn the instance on the hypervisor. [ 989.747891] env[62183]: DEBUG nova.compute.manager [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 989.751503] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1018fd3-0403-443c-bb36-7d6f4429f340 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.814657] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f4b750-2287-08e5-ab8e-ee9079402a3c, 'name': SearchDatastore_Task, 'duration_secs': 0.013224} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.815500] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a83a607-edbe-4a7b-ba53-13e3dfddbc2e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.821964] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 989.821964] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525367a5-424c-b1ac-6d6c-ed34886c4527" [ 989.821964] env[62183]: _type = "Task" [ 989.821964] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.830564] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525367a5-424c-b1ac-6d6c-ed34886c4527, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.057017] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "dbd34888-a333-40b4-ae9c-df5541b16704" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.057017] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "dbd34888-a333-40b4-ae9c-df5541b16704" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.134936] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387361, 'name': Rename_Task, 'duration_secs': 0.362607} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.137134] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.137134] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-824c72e6-fa18-499d-9f93-9bf19cb8a935 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.144418] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 990.144418] env[62183]: value = "task-1387362" [ 990.144418] env[62183]: _type = "Task" [ 990.144418] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.152615] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387362, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.233470] env[62183]: DEBUG nova.compute.manager [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 990.270686] env[62183]: INFO nova.compute.manager [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Took 18.02 seconds to build instance. [ 990.335501] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525367a5-424c-b1ac-6d6c-ed34886c4527, 'name': SearchDatastore_Task, 'duration_secs': 0.011333} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.337472] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.337745] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4/d8817d24-b0cb-4956-b195-cc417ae09fb4.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.338235] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-791e8473-d247-4f3b-8831-c26b96c86c8c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.346283] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 990.346283] env[62183]: value = "task-1387363" [ 990.346283] env[62183]: _type = "Task" [ 990.346283] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.356695] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387363, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.375613] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.375870] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.376692] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.376692] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.376692] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.379094] env[62183]: INFO nova.compute.manager [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Terminating instance [ 990.381960] env[62183]: DEBUG nova.compute.manager [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 990.381960] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 990.382519] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8afa42-e9de-465a-b1a6-0e20691b3d91 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.391920] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 990.391920] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-339d5f4b-9579-4f13-b065-58f94bf556a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.403280] env[62183]: DEBUG oslo_vmware.api [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 990.403280] env[62183]: value = "task-1387364" [ 990.403280] env[62183]: _type = "Task" [ 990.403280] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.413850] env[62183]: DEBUG oslo_vmware.api [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387364, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.445622] env[62183]: DEBUG oslo_concurrency.lockutils [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "143323db-da1c-4dd8-8f13-ed0af31e2027" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.445955] env[62183]: DEBUG oslo_concurrency.lockutils [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "143323db-da1c-4dd8-8f13-ed0af31e2027" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.446240] env[62183]: DEBUG oslo_concurrency.lockutils [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "143323db-da1c-4dd8-8f13-ed0af31e2027-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.446540] env[62183]: DEBUG oslo_concurrency.lockutils [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "143323db-da1c-4dd8-8f13-ed0af31e2027-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.446752] env[62183]: DEBUG oslo_concurrency.lockutils [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "143323db-da1c-4dd8-8f13-ed0af31e2027-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.452341] env[62183]: INFO nova.compute.manager [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Terminating instance [ 990.455926] env[62183]: DEBUG nova.compute.manager [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 990.456207] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 990.457138] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b771a99e-cd8b-4e35-8574-ecce13babfab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.467674] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 990.471151] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b02e344-74eb-4d0b-823c-29cad486cefa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.480045] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.480379] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.486222] env[62183]: DEBUG oslo_vmware.api [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 990.486222] env[62183]: value = "task-1387365" [ 990.486222] env[62183]: _type = "Task" [ 990.486222] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.491335] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30c6af9-c694-42af-a308-bf7bb551b96a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.502804] env[62183]: DEBUG oslo_vmware.api [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387365, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.503922] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a891436f-ac89-4931-a548-30b7a0b44336 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.544137] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32752646-3255-4935-b998-b19126f89807 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.554547] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8f77ea-6b63-4e9c-9eb2-a162eda3ddec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.559338] env[62183]: DEBUG nova.compute.manager [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 990.574203] env[62183]: DEBUG nova.compute.provider_tree [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.659932] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387362, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.757261] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.773164] env[62183]: DEBUG oslo_concurrency.lockutils [None req-44b4a72f-5e72-43d8-8c93-724c880d9708 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "036037b3-9676-419c-ab87-e5a557438fbe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.535s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.858158] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387363, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.913859] env[62183]: DEBUG oslo_vmware.api [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387364, 'name': PowerOffVM_Task, 'duration_secs': 0.196601} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.914165] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 990.914351] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 990.914618] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6f3adb6-0702-4aa8-8172-7b877bf13ba4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.982915] env[62183]: DEBUG nova.compute.manager [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 990.993877] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.994114] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.994207] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleting the datastore file [datastore1] b6402e2f-1035-4c1b-9a1e-05b17d89e4bb {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.994557] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b13bb5fd-7b11-4d7b-8fb5-a2d69b648280 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.003523] env[62183]: DEBUG oslo_vmware.api [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387365, 'name': PowerOffVM_Task, 'duration_secs': 0.21406} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.004030] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 991.004212] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 991.004506] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14365617-017a-483d-bc5a-babeae67bb93 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.009026] env[62183]: DEBUG oslo_vmware.api [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 991.009026] env[62183]: value = "task-1387367" [ 991.009026] env[62183]: _type = "Task" [ 991.009026] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.017702] env[62183]: DEBUG oslo_vmware.api [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387367, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.080023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.080944] env[62183]: DEBUG nova.scheduler.client.report [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 991.085412] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 991.085632] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 991.085887] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleting the datastore file [datastore1] 143323db-da1c-4dd8-8f13-ed0af31e2027 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.087425] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d85e7569-5d9e-4623-b360-4b82670ec728 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.098766] env[62183]: DEBUG oslo_vmware.api [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for the task: (returnval){ [ 991.098766] env[62183]: value = "task-1387369" [ 991.098766] env[62183]: _type = "Task" [ 991.098766] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.108371] env[62183]: DEBUG oslo_vmware.api [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387369, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.155646] env[62183]: DEBUG oslo_vmware.api [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387362, 'name': PowerOnVM_Task, 'duration_secs': 0.598426} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.155932] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 991.156155] env[62183]: INFO nova.compute.manager [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Took 7.98 seconds to spawn the instance on the hypervisor. [ 991.156340] env[62183]: DEBUG nova.compute.manager [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.157657] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609024d8-ece9-4189-b4a8-009322b2dc7c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.160674] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.160908] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.161199] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.161391] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.161564] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.167769] env[62183]: INFO nova.compute.manager [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Terminating instance [ 991.169425] env[62183]: DEBUG nova.compute.manager [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 991.169425] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 991.170137] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a223ad9-8d78-4227-b579-7f5e02cedd95 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.178626] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.178958] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31f3aaf3-d7a6-4999-8594-cbb6ddab9043 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.186026] env[62183]: DEBUG oslo_vmware.api [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 991.186026] env[62183]: value = "task-1387370" [ 991.186026] env[62183]: _type = "Task" [ 991.186026] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.194501] env[62183]: DEBUG oslo_vmware.api [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387370, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.358299] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387363, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669466} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.358532] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4/d8817d24-b0cb-4956-b195-cc417ae09fb4.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.358755] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.359023] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b34cef8-46de-4142-b5ac-a2b28eee9f9e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.365559] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 991.365559] env[62183]: value = "task-1387371" [ 991.365559] env[62183]: _type = "Task" [ 991.365559] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.374272] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.505160] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.518936] env[62183]: DEBUG oslo_vmware.api [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387367, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.299176} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.519228] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 991.519625] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 991.519625] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 991.519749] env[62183]: INFO nova.compute.manager [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Took 1.14 seconds to destroy the instance on the hypervisor. [ 991.519994] env[62183]: DEBUG oslo.service.loopingcall [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.520207] env[62183]: DEBUG nova.compute.manager [-] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 991.520300] env[62183]: DEBUG nova.network.neutron [-] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 991.592741] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquiring lock "036037b3-9676-419c-ab87-e5a557438fbe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.593014] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "036037b3-9676-419c-ab87-e5a557438fbe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.593244] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquiring lock "036037b3-9676-419c-ab87-e5a557438fbe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.593446] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "036037b3-9676-419c-ab87-e5a557438fbe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.593655] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "036037b3-9676-419c-ab87-e5a557438fbe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.595785] env[62183]: INFO nova.compute.manager [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Terminating instance [ 991.599049] env[62183]: DEBUG nova.compute.manager [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 991.599288] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 991.600227] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09dae53c-5289-4d08-9e6f-e70e62bd477a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.612963] env[62183]: DEBUG oslo_vmware.api [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Task: {'id': task-1387369, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259032} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.615117] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 991.615314] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 991.615495] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 991.615666] env[62183]: INFO nova.compute.manager [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Took 1.16 seconds to destroy the instance on the hypervisor. [ 991.615903] env[62183]: DEBUG oslo.service.loopingcall [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.616172] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.616403] env[62183]: DEBUG nova.compute.manager [-] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 991.616488] env[62183]: DEBUG nova.network.neutron [-] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 991.618000] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af3c515e-e6d0-42f6-8cc3-35b7016af90b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.625284] env[62183]: DEBUG oslo_vmware.api [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 991.625284] env[62183]: value = "task-1387372" [ 991.625284] env[62183]: _type = "Task" [ 991.625284] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.635652] env[62183]: DEBUG oslo_vmware.api [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387372, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.677309] env[62183]: INFO nova.compute.manager [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Took 19.09 seconds to build instance. [ 991.700842] env[62183]: DEBUG oslo_vmware.api [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387370, 'name': PowerOffVM_Task, 'duration_secs': 0.357503} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.700842] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 991.700842] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 991.700842] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e1a25eb-4622-4b73-a30c-deac423bcee9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.812024] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 991.812024] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 991.812024] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleting the datastore file [datastore1] c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.812024] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-090d776c-e881-40c3-8dd2-a6d6033ce4ed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.818141] env[62183]: DEBUG oslo_vmware.api [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 991.818141] env[62183]: value = "task-1387374" [ 991.818141] env[62183]: _type = "Task" [ 991.818141] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.830098] env[62183]: DEBUG oslo_vmware.api [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387374, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.875309] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07977} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.876048] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.876876] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f24909-64db-4108-b459-df335be9d154 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.900149] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4/d8817d24-b0cb-4956-b195-cc417ae09fb4.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.900475] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb640368-cf2e-485f-b547-7b28312a3693 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.917935] env[62183]: DEBUG nova.compute.manager [req-ef2a4ad4-ac0d-4f61-add9-7f9e5ab26334 req-0c1df161-3759-498b-9b04-57d1adb0e144 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Received event network-vif-deleted-9cc46a34-db54-4e5b-97fe-e3b4968da907 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.918152] env[62183]: INFO nova.compute.manager [req-ef2a4ad4-ac0d-4f61-add9-7f9e5ab26334 req-0c1df161-3759-498b-9b04-57d1adb0e144 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Neutron deleted interface 9cc46a34-db54-4e5b-97fe-e3b4968da907; detaching it from the instance and deleting it from the info cache [ 991.918332] env[62183]: DEBUG nova.network.neutron [req-ef2a4ad4-ac0d-4f61-add9-7f9e5ab26334 req-0c1df161-3759-498b-9b04-57d1adb0e144 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.925050] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 991.925050] env[62183]: value = "task-1387375" [ 991.925050] env[62183]: _type = "Task" [ 991.925050] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.935965] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387375, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.938839] env[62183]: DEBUG nova.compute.manager [req-674a5532-3617-4cc2-a08c-fe3021fcd247 req-226ba5b4-427b-47b3-a069-ccafd16866d0 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Received event network-vif-deleted-81a4f8ba-127a-4d45-9fe2-462656e5ff9d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.938839] env[62183]: INFO nova.compute.manager [req-674a5532-3617-4cc2-a08c-fe3021fcd247 req-226ba5b4-427b-47b3-a069-ccafd16866d0 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Neutron deleted interface 81a4f8ba-127a-4d45-9fe2-462656e5ff9d; detaching it from the instance and deleting it from the info cache [ 991.938839] env[62183]: DEBUG nova.network.neutron [req-674a5532-3617-4cc2-a08c-fe3021fcd247 req-226ba5b4-427b-47b3-a069-ccafd16866d0 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.093014] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.381s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.093014] env[62183]: DEBUG nova.compute.manager [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62183) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 992.095708] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.339s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.097155] env[62183]: INFO nova.compute.claims [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 992.136145] env[62183]: DEBUG oslo_vmware.api [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387372, 'name': PowerOffVM_Task, 'duration_secs': 0.392522} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.137452] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.137734] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 992.138035] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d56877b-8cd6-4f80-b842-1a4fce230170 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.179211] env[62183]: DEBUG oslo_concurrency.lockutils [None req-168279e4-c087-4f5f-93d9-276c9826f7d9 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.596s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.269739] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 992.269739] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 992.269739] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Deleting the datastore file [datastore1] 036037b3-9676-419c-ab87-e5a557438fbe {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.269739] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a40671fe-9eda-42b0-80f3-983e5fab57df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.277460] env[62183]: DEBUG oslo_vmware.api [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for the task: (returnval){ [ 992.277460] env[62183]: value = "task-1387377" [ 992.277460] env[62183]: _type = "Task" [ 992.277460] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.287031] env[62183]: DEBUG oslo_vmware.api [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387377, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.330556] env[62183]: DEBUG oslo_vmware.api [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387374, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203197} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.330556] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.330556] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.330556] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 992.330556] env[62183]: INFO nova.compute.manager [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 992.330556] env[62183]: DEBUG oslo.service.loopingcall [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.330556] env[62183]: DEBUG nova.compute.manager [-] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.330556] env[62183]: DEBUG nova.network.neutron [-] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.339276] env[62183]: DEBUG nova.network.neutron [-] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.376373] env[62183]: DEBUG nova.network.neutron [-] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.420978] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-306ed0e9-7f26-48e6-bb07-7e2f84623f11 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.433967] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d691be-d34a-4588-b1eb-83f668468505 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.449448] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387375, 'name': ReconfigVM_Task, 'duration_secs': 0.315179} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.449708] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-854ea3c0-cb8a-4dd9-a28c-6d6606c9189c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.451919] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Reconfigured VM instance instance-00000061 to attach disk [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4/d8817d24-b0cb-4956-b195-cc417ae09fb4.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.452619] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38438c22-1fd9-4d85-81e4-d762b26f45e7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.472305] env[62183]: DEBUG nova.compute.manager [req-ef2a4ad4-ac0d-4f61-add9-7f9e5ab26334 req-0c1df161-3759-498b-9b04-57d1adb0e144 service nova] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Detach interface failed, port_id=9cc46a34-db54-4e5b-97fe-e3b4968da907, reason: Instance b6402e2f-1035-4c1b-9a1e-05b17d89e4bb could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 992.474846] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 992.474846] env[62183]: value = "task-1387378" [ 992.474846] env[62183]: _type = "Task" [ 992.474846] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.477846] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c2773a-cd0b-49ec-8c5b-b39f4ffe51d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.501239] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387378, 'name': Rename_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.511741] env[62183]: DEBUG nova.compute.manager [req-674a5532-3617-4cc2-a08c-fe3021fcd247 req-226ba5b4-427b-47b3-a069-ccafd16866d0 service nova] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Detach interface failed, port_id=81a4f8ba-127a-4d45-9fe2-462656e5ff9d, reason: Instance 143323db-da1c-4dd8-8f13-ed0af31e2027 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 992.657786] env[62183]: INFO nova.scheduler.client.report [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted allocation for migration 4e58e303-0d2b-4df3-96ab-52a8e649d12b [ 992.717735] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.717983] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.718211] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.718395] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.718563] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.720531] env[62183]: INFO nova.compute.manager [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Terminating instance [ 992.723073] env[62183]: DEBUG nova.compute.manager [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 992.723269] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 992.724107] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862d30e6-af8b-451f-81f1-573da4225b87 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.732326] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.732569] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed47b58c-5ee4-41c9-92fa-4e645676aa16 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.739431] env[62183]: DEBUG oslo_vmware.api [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 992.739431] env[62183]: value = "task-1387379" [ 992.739431] env[62183]: _type = "Task" [ 992.739431] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.747688] env[62183]: DEBUG oslo_vmware.api [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387379, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.788836] env[62183]: DEBUG oslo_vmware.api [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Task: {'id': task-1387377, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190115} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.789077] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.789297] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 992.789489] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 992.789682] env[62183]: INFO nova.compute.manager [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Took 1.19 seconds to destroy the instance on the hypervisor. [ 992.789944] env[62183]: DEBUG oslo.service.loopingcall [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.790190] env[62183]: DEBUG nova.compute.manager [-] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 992.790315] env[62183]: DEBUG nova.network.neutron [-] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.842574] env[62183]: INFO nova.compute.manager [-] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Took 1.32 seconds to deallocate network for instance. [ 992.879438] env[62183]: INFO nova.compute.manager [-] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Took 1.26 seconds to deallocate network for instance. [ 992.987295] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387378, 'name': Rename_Task, 'duration_secs': 0.1441} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.987682] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.988012] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c7f5afe-d738-427e-8bca-b57d80f3daf4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.995911] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 992.995911] env[62183]: value = "task-1387380" [ 992.995911] env[62183]: _type = "Task" [ 992.995911] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.009296] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387380, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.166023] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4527b8f8-5349-4f88-96c9-6cd2be51d294 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.802s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.253052] env[62183]: DEBUG oslo_vmware.api [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387379, 'name': PowerOffVM_Task, 'duration_secs': 0.251127} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.253560] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.253715] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 993.253985] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48f38d4e-86f9-4082-8105-24af195288cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.332232] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80c6d7d-1c59-4320-8ad4-3913fdff601c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.334950] env[62183]: DEBUG nova.network.neutron [-] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.342043] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe364e10-bcc5-4c9f-95c2-121dafe2e452 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.348977] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.375677] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e12f661-481b-4bdc-922f-a0089bc1d3cb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.390034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.390034] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59283f52-36cc-452a-8edd-154e695946cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.402763] env[62183]: DEBUG nova.compute.provider_tree [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.505850] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387380, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.592848] env[62183]: DEBUG nova.network.neutron [-] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.666815] env[62183]: DEBUG nova.objects.instance [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'flavor' on Instance uuid a948464b-63aa-4bc8-9885-228049e96d37 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.837822] env[62183]: INFO nova.compute.manager [-] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Took 1.51 seconds to deallocate network for instance. [ 993.883179] env[62183]: DEBUG oslo_concurrency.lockutils [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.883570] env[62183]: DEBUG oslo_concurrency.lockutils [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.883811] env[62183]: DEBUG oslo_concurrency.lockutils [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.884026] env[62183]: DEBUG oslo_concurrency.lockutils [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.884211] env[62183]: DEBUG oslo_concurrency.lockutils [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.886263] env[62183]: INFO nova.compute.manager [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Terminating instance [ 993.888376] env[62183]: DEBUG nova.compute.manager [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 993.888493] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 993.889605] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e6ef4d-1d69-48c8-91fa-60fce280ff30 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.898060] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 993.898296] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e5f8ed5-e181-450d-807c-926a4a859541 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.904295] env[62183]: DEBUG oslo_vmware.api [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 993.904295] env[62183]: value = "task-1387382" [ 993.904295] env[62183]: _type = "Task" [ 993.904295] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.908150] env[62183]: DEBUG nova.scheduler.client.report [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 993.916551] env[62183]: DEBUG oslo_vmware.api [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.943086] env[62183]: DEBUG nova.compute.manager [req-62e86a52-8f3e-4e5d-ae05-682614b30442 req-4ba0e6a5-b27d-413d-b95b-fbd040d0e4d2 service nova] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Received event network-vif-deleted-f9702809-bb25-463e-a193-580acaeaf306 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.964198] env[62183]: DEBUG nova.compute.manager [req-163b09a6-467c-4bf3-81a7-88b52b7b471f req-de7eda8d-4e60-4f41-ad9e-0b4aa740d363 service nova] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Received event network-vif-deleted-31f6cef9-f14a-4001-a29d-4f02661bbf2d {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.006669] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387380, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.095020] env[62183]: INFO nova.compute.manager [-] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Took 1.30 seconds to deallocate network for instance. [ 994.171408] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.171601] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.171785] env[62183]: DEBUG nova.network.neutron [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 994.171964] env[62183]: DEBUG nova.objects.instance [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'info_cache' on Instance uuid a948464b-63aa-4bc8-9885-228049e96d37 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.344166] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.414980] env[62183]: DEBUG oslo_vmware.api [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.417887] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.418389] env[62183]: DEBUG nova.compute.manager [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 994.421260] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.341s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.422601] env[62183]: INFO nova.compute.claims [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.507610] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387380, 'name': PowerOnVM_Task} progress is 70%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.573713] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 994.574279] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 994.574279] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleting the datastore file [datastore1] 24f7c072-1209-4c6e-9a54-4e069f14f7d2 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.574766] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a5e70ace-52ce-47aa-8e74-41a1806cc9a4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.582624] env[62183]: DEBUG oslo_vmware.api [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 994.582624] env[62183]: value = "task-1387383" [ 994.582624] env[62183]: _type = "Task" [ 994.582624] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.593120] env[62183]: DEBUG oslo_vmware.api [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387383, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.601349] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.676043] env[62183]: DEBUG nova.objects.base [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 994.915642] env[62183]: DEBUG oslo_vmware.api [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387382, 'name': PowerOffVM_Task, 'duration_secs': 0.705393} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.915996] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 994.916238] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 994.916539] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82975b62-7d85-44bd-8242-afe5fafdb41d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.922513] env[62183]: DEBUG nova.compute.utils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 994.923867] env[62183]: DEBUG nova.compute.manager [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 994.924528] env[62183]: DEBUG nova.network.neutron [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 994.965420] env[62183]: DEBUG nova.policy [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4f1ba64ecf948ddabab1c59745f9fc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '587870b99b844c819d27ac4852e44c3d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 995.008796] env[62183]: DEBUG oslo_vmware.api [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387380, 'name': PowerOnVM_Task, 'duration_secs': 1.996703} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.009077] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.009300] env[62183]: INFO nova.compute.manager [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Took 9.51 seconds to spawn the instance on the hypervisor. [ 995.009488] env[62183]: DEBUG nova.compute.manager [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.010306] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c6c1a5-429c-46e5-be98-109af0b675d2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.093339] env[62183]: DEBUG oslo_vmware.api [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387383, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165954} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.093589] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.095022] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.095022] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.095022] env[62183]: INFO nova.compute.manager [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Took 2.37 seconds to destroy the instance on the hypervisor. [ 995.095022] env[62183]: DEBUG oslo.service.loopingcall [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.095022] env[62183]: DEBUG nova.compute.manager [-] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 995.095022] env[62183]: DEBUG nova.network.neutron [-] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.262739] env[62183]: DEBUG nova.network.neutron [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Successfully created port: f48d3c73-edab-4cf3-94cc-498503b5c465 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 995.417907] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 995.418200] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 995.418329] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Deleting the datastore file [datastore1] fb348784-62a0-4d1f-ac7f-f176f3da0dd9 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.418609] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de5319d6-650c-449b-9280-88029c63ab47 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.425312] env[62183]: DEBUG oslo_vmware.api [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for the task: (returnval){ [ 995.425312] env[62183]: value = "task-1387385" [ 995.425312] env[62183]: _type = "Task" [ 995.425312] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.428987] env[62183]: DEBUG nova.compute.manager [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 995.444219] env[62183]: DEBUG oslo_vmware.api [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.535802] env[62183]: INFO nova.compute.manager [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Took 21.19 seconds to build instance. [ 996.229312] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3d6ef7d1-d6c3-454c-8e56-be992d1e266b tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "d8817d24-b0cb-4956-b195-cc417ae09fb4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.887s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.238157] env[62183]: DEBUG nova.compute.manager [req-570387aa-0f19-4dfe-9f6e-c5ceb8030204 req-366082d3-3cc6-4d24-8247-19a864f6f073 service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Received event network-vif-deleted-1fa5a463-ffcd-41f4-b61d-3fa88e6178c9 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.238157] env[62183]: INFO nova.compute.manager [req-570387aa-0f19-4dfe-9f6e-c5ceb8030204 req-366082d3-3cc6-4d24-8247-19a864f6f073 service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Neutron deleted interface 1fa5a463-ffcd-41f4-b61d-3fa88e6178c9; detaching it from the instance and deleting it from the info cache [ 996.238157] env[62183]: DEBUG nova.network.neutron [req-570387aa-0f19-4dfe-9f6e-c5ceb8030204 req-366082d3-3cc6-4d24-8247-19a864f6f073 service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.245920] env[62183]: DEBUG oslo_vmware.api [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Task: {'id': task-1387385, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126132} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.246180] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.246361] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 996.246563] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 996.246705] env[62183]: INFO nova.compute.manager [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Took 2.36 seconds to destroy the instance on the hypervisor. [ 996.246944] env[62183]: DEBUG oslo.service.loopingcall [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.247372] env[62183]: DEBUG nova.compute.manager [-] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 996.247472] env[62183]: DEBUG nova.network.neutron [-] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.346215] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f43ad43-765e-4715-9c55-593e9d8ac14a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.357019] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0454d3ef-d07d-4cc4-be99-02c92370ff2d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.383235] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e784ba-d1b3-4cd7-a6e3-8ac0dbc1eafa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.390488] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035a80d5-6d46-4eb7-9f8c-5018c8f7f48d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.405856] env[62183]: DEBUG nova.compute.provider_tree [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.480665] env[62183]: DEBUG nova.network.neutron [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [{"id": "3d687443-41fd-44ff-8981-35319b79d88a", "address": "fa:16:3e:91:31:9b", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d687443-41", "ovs_interfaceid": "3d687443-41fd-44ff-8981-35319b79d88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.731721] env[62183]: DEBUG nova.compute.manager [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 996.741905] env[62183]: DEBUG nova.network.neutron [-] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.741905] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-39ccadbc-76c7-4007-9fd5-3ec8dd811990 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.752866] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f6b9f1-b5d7-4f2f-abc8-3409c2cec03a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.765633] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 996.765875] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 996.766044] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 996.766323] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 996.766490] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 996.766684] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 996.766890] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 996.767062] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 996.767235] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 996.767398] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 996.767572] env[62183]: DEBUG nova.virt.hardware [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 996.768612] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbef5d4-1188-4225-ba53-5b168e622be7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.779652] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c6ac11-9302-4510-a5a8-6b824fdb8296 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.794475] env[62183]: DEBUG nova.compute.manager [req-570387aa-0f19-4dfe-9f6e-c5ceb8030204 req-366082d3-3cc6-4d24-8247-19a864f6f073 service nova] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Detach interface failed, port_id=1fa5a463-ffcd-41f4-b61d-3fa88e6178c9, reason: Instance 24f7c072-1209-4c6e-9a54-4e069f14f7d2 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 996.796885] env[62183]: DEBUG nova.compute.manager [req-e0a6fe2a-6b48-497c-95a1-9451d78241ae req-25fbbb09-1345-4314-940d-1315dea62d1a service nova] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Received event network-vif-plugged-f48d3c73-edab-4cf3-94cc-498503b5c465 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.797103] env[62183]: DEBUG oslo_concurrency.lockutils [req-e0a6fe2a-6b48-497c-95a1-9451d78241ae req-25fbbb09-1345-4314-940d-1315dea62d1a service nova] Acquiring lock "a5018b4e-016d-4c61-bfbf-2c039ab79499-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.798305] env[62183]: DEBUG oslo_concurrency.lockutils [req-e0a6fe2a-6b48-497c-95a1-9451d78241ae req-25fbbb09-1345-4314-940d-1315dea62d1a service nova] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.798305] env[62183]: DEBUG oslo_concurrency.lockutils [req-e0a6fe2a-6b48-497c-95a1-9451d78241ae req-25fbbb09-1345-4314-940d-1315dea62d1a service nova] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.798305] env[62183]: DEBUG nova.compute.manager [req-e0a6fe2a-6b48-497c-95a1-9451d78241ae req-25fbbb09-1345-4314-940d-1315dea62d1a service nova] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] No waiting events found dispatching network-vif-plugged-f48d3c73-edab-4cf3-94cc-498503b5c465 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 996.798305] env[62183]: WARNING nova.compute.manager [req-e0a6fe2a-6b48-497c-95a1-9451d78241ae req-25fbbb09-1345-4314-940d-1315dea62d1a service nova] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Received unexpected event network-vif-plugged-f48d3c73-edab-4cf3-94cc-498503b5c465 for instance with vm_state building and task_state spawning. [ 996.873481] env[62183]: DEBUG nova.network.neutron [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Successfully updated port: f48d3c73-edab-4cf3-94cc-498503b5c465 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 996.909588] env[62183]: DEBUG nova.scheduler.client.report [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 996.982992] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-a948464b-63aa-4bc8-9885-228049e96d37" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.050754] env[62183]: INFO nova.compute.manager [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Rebuilding instance [ 997.096905] env[62183]: DEBUG nova.compute.manager [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.097783] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53865f6f-80d6-44a6-a5cb-ca235cfde1d1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.241955] env[62183]: INFO nova.compute.manager [-] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Took 2.15 seconds to deallocate network for instance. [ 997.324703] env[62183]: DEBUG nova.network.neutron [-] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.376207] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "refresh_cache-a5018b4e-016d-4c61-bfbf-2c039ab79499" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.376362] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "refresh_cache-a5018b4e-016d-4c61-bfbf-2c039ab79499" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.376510] env[62183]: DEBUG nova.network.neutron [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 997.415032] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.993s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.415451] env[62183]: DEBUG nova.compute.manager [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 997.418285] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.913s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.419627] env[62183]: INFO nova.compute.claims [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 997.486015] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.486697] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1bc768a5-a5ee-49d9-9305-6faa60393798 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.496763] env[62183]: DEBUG oslo_vmware.api [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 997.496763] env[62183]: value = "task-1387386" [ 997.496763] env[62183]: _type = "Task" [ 997.496763] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.505050] env[62183]: DEBUG oslo_vmware.api [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387386, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.608998] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.609365] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b5b907c-9453-4105-adb6-56647460c827 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.616322] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 997.616322] env[62183]: value = "task-1387387" [ 997.616322] env[62183]: _type = "Task" [ 997.616322] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.625895] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387387, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.748268] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.828136] env[62183]: INFO nova.compute.manager [-] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Took 1.58 seconds to deallocate network for instance. [ 997.909809] env[62183]: DEBUG nova.network.neutron [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 997.927073] env[62183]: DEBUG nova.compute.utils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.931968] env[62183]: DEBUG nova.compute.manager [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 997.931968] env[62183]: DEBUG nova.network.neutron [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 997.970422] env[62183]: DEBUG nova.policy [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4f1ba64ecf948ddabab1c59745f9fc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '587870b99b844c819d27ac4852e44c3d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 998.007564] env[62183]: DEBUG oslo_vmware.api [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387386, 'name': PowerOnVM_Task, 'duration_secs': 0.37388} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.007564] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 998.007747] env[62183]: DEBUG nova.compute.manager [None req-e586b627-80d5-4779-a5d0-42ddd3b18b8f tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.008762] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac20850d-e225-4e15-b1f1-0ec5c88aa246 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.026407] env[62183]: DEBUG nova.compute.manager [req-ea1d66e3-7e9e-4a1f-9fe3-0616cb975750 req-2ee8f629-e6c6-4b34-b67c-56b98b603c8e service nova] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Received event network-vif-deleted-37fafdd3-7651-4cfc-818d-2ea8e63ffcb1 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.059145] env[62183]: DEBUG nova.network.neutron [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Updating instance_info_cache with network_info: [{"id": "f48d3c73-edab-4cf3-94cc-498503b5c465", "address": "fa:16:3e:65:a7:94", "network": {"id": "7cbe5832-931e-430d-95f5-8f48781f095b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-236543802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "587870b99b844c819d27ac4852e44c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf48d3c73-ed", "ovs_interfaceid": "f48d3c73-edab-4cf3-94cc-498503b5c465", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.125792] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387387, 'name': PowerOffVM_Task, 'duration_secs': 0.168557} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.126065] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.126291] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 998.127132] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a71d1f8-d701-4c60-942a-a54eba442cd8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.134700] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.134994] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af57c3db-a2ce-4927-8f3b-435f0c2fc313 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.202196] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.202530] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.202739] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleting the datastore file [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.203618] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bbb90cc3-dde9-49cb-aee9-6902d3c6a9c4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.209890] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 998.209890] env[62183]: value = "task-1387389" [ 998.209890] env[62183]: _type = "Task" [ 998.209890] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.218416] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387389, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.221015] env[62183]: DEBUG nova.network.neutron [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Successfully created port: b7e8fcf5-0b5e-4b35-af6b-f18163968d76 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 998.337689] env[62183]: DEBUG oslo_concurrency.lockutils [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.434657] env[62183]: DEBUG nova.compute.manager [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 998.561108] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "refresh_cache-a5018b4e-016d-4c61-bfbf-2c039ab79499" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.561764] env[62183]: DEBUG nova.compute.manager [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Instance network_info: |[{"id": "f48d3c73-edab-4cf3-94cc-498503b5c465", "address": "fa:16:3e:65:a7:94", "network": {"id": "7cbe5832-931e-430d-95f5-8f48781f095b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-236543802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "587870b99b844c819d27ac4852e44c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf48d3c73-ed", "ovs_interfaceid": "f48d3c73-edab-4cf3-94cc-498503b5c465", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 998.561893] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:a7:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e55c248-c504-4c7a-bbe9-f42cf417aee7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f48d3c73-edab-4cf3-94cc-498503b5c465', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 998.569445] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Creating folder: Project (587870b99b844c819d27ac4852e44c3d). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 998.572563] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0df1c74e-e4ed-45d3-bbac-237275fceb3c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.582698] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Created folder: Project (587870b99b844c819d27ac4852e44c3d) in parent group-v294392. [ 998.582915] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Creating folder: Instances. Parent ref: group-v294507. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 998.583200] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b279065c-8fe2-4f36-9c65-f500d7c77e74 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.594205] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Created folder: Instances in parent group-v294507. [ 998.594494] env[62183]: DEBUG oslo.service.loopingcall [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.594705] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 998.594935] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea3016fa-ee39-48fa-b35d-e0028e393457 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.615963] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 998.615963] env[62183]: value = "task-1387392" [ 998.615963] env[62183]: _type = "Task" [ 998.615963] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.623255] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387392, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.667474] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba57d171-3add-4d91-a698-72acba5a8f64 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.674209] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ebbe33-f537-4cd0-a2fd-06396a8bce4f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.705431] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac7329f-32c4-428f-9414-fc28b803a8a5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.715708] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33b0d0d-927c-4a8d-9d9d-2455c8335c65 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.723412] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387389, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134055} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.732475] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 998.732671] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 998.732849] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 998.735516] env[62183]: DEBUG nova.compute.provider_tree [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.869857] env[62183]: DEBUG nova.compute.manager [req-86f4a87d-007c-4568-b205-db667845e5a9 req-37892996-5fbf-45c4-bdf5-c06f49ad15a4 service nova] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Received event network-changed-f48d3c73-edab-4cf3-94cc-498503b5c465 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.870081] env[62183]: DEBUG nova.compute.manager [req-86f4a87d-007c-4568-b205-db667845e5a9 req-37892996-5fbf-45c4-bdf5-c06f49ad15a4 service nova] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Refreshing instance network info cache due to event network-changed-f48d3c73-edab-4cf3-94cc-498503b5c465. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 998.870324] env[62183]: DEBUG oslo_concurrency.lockutils [req-86f4a87d-007c-4568-b205-db667845e5a9 req-37892996-5fbf-45c4-bdf5-c06f49ad15a4 service nova] Acquiring lock "refresh_cache-a5018b4e-016d-4c61-bfbf-2c039ab79499" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.870487] env[62183]: DEBUG oslo_concurrency.lockutils [req-86f4a87d-007c-4568-b205-db667845e5a9 req-37892996-5fbf-45c4-bdf5-c06f49ad15a4 service nova] Acquired lock "refresh_cache-a5018b4e-016d-4c61-bfbf-2c039ab79499" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.870767] env[62183]: DEBUG nova.network.neutron [req-86f4a87d-007c-4568-b205-db667845e5a9 req-37892996-5fbf-45c4-bdf5-c06f49ad15a4 service nova] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Refreshing network info cache for port f48d3c73-edab-4cf3-94cc-498503b5c465 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.126700] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387392, 'name': CreateVM_Task} progress is 25%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.241762] env[62183]: DEBUG nova.scheduler.client.report [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 999.443754] env[62183]: DEBUG nova.compute.manager [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 999.470667] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.470929] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.471104] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.471298] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.471450] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.471602] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.471815] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.471978] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.472167] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.472333] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.472503] env[62183]: DEBUG nova.virt.hardware [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.473362] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16bc86a-762e-4e2d-96d3-a228bdb75324 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.481251] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a248a6c-5a39-48d3-bb5d-2c8f07cf4bc5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.570712] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "a948464b-63aa-4bc8-9885-228049e96d37" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.570712] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.570712] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "a948464b-63aa-4bc8-9885-228049e96d37-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.570712] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.570712] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.572668] env[62183]: INFO nova.compute.manager [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Terminating instance [ 999.574505] env[62183]: DEBUG nova.compute.manager [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.574739] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.575596] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6712ac-3689-4bd8-89b9-e2a881238d6b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.579030] env[62183]: DEBUG nova.network.neutron [req-86f4a87d-007c-4568-b205-db667845e5a9 req-37892996-5fbf-45c4-bdf5-c06f49ad15a4 service nova] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Updated VIF entry in instance network info cache for port f48d3c73-edab-4cf3-94cc-498503b5c465. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 999.579382] env[62183]: DEBUG nova.network.neutron [req-86f4a87d-007c-4568-b205-db667845e5a9 req-37892996-5fbf-45c4-bdf5-c06f49ad15a4 service nova] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Updating instance_info_cache with network_info: [{"id": "f48d3c73-edab-4cf3-94cc-498503b5c465", "address": "fa:16:3e:65:a7:94", "network": {"id": "7cbe5832-931e-430d-95f5-8f48781f095b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-236543802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "587870b99b844c819d27ac4852e44c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf48d3c73-ed", "ovs_interfaceid": "f48d3c73-edab-4cf3-94cc-498503b5c465", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.586077] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.586186] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d5939b7-d6ee-4c08-88d5-cc8c56ecd1df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.592758] env[62183]: DEBUG oslo_vmware.api [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 999.592758] env[62183]: value = "task-1387393" [ 999.592758] env[62183]: _type = "Task" [ 999.592758] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.603016] env[62183]: DEBUG oslo_vmware.api [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.625990] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387392, 'name': CreateVM_Task, 'duration_secs': 0.683571} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.626191] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 999.626932] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.627127] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.627487] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 999.628054] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc378e87-ed64-4732-bfab-9e13eecccd01 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.632349] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 999.632349] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525806a6-0eaf-4264-791d-3cbc426cc864" [ 999.632349] env[62183]: _type = "Task" [ 999.632349] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.640278] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525806a6-0eaf-4264-791d-3cbc426cc864, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.716730] env[62183]: DEBUG nova.network.neutron [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Successfully updated port: b7e8fcf5-0b5e-4b35-af6b-f18163968d76 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.749958] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.750575] env[62183]: DEBUG nova.compute.manager [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 999.753950] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.405s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.754198] env[62183]: DEBUG nova.objects.instance [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lazy-loading 'resources' on Instance uuid b6402e2f-1035-4c1b-9a1e-05b17d89e4bb {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.773620] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.773865] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.774039] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.774237] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.774388] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.774568] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.774910] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.775102] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.775306] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.775508] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.775706] env[62183]: DEBUG nova.virt.hardware [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.776933] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b627cd-2e3f-464a-8a8f-7d0aa578836f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.785454] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56d3199-8c30-452d-a053-bd4e994646f2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.799598] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:dd:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8800a981-a89e-42e4-8be9-cace419ba9cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b2e8106-ed68-4023-b3cc-d91842d269be', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.806876] env[62183]: DEBUG oslo.service.loopingcall [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.807355] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 999.807570] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0293d09-e8d1-46e1-8742-3a472dda1e13 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.825659] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.825659] env[62183]: value = "task-1387394" [ 999.825659] env[62183]: _type = "Task" [ 999.825659] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.832674] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387394, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.058601] env[62183]: DEBUG nova.compute.manager [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Received event network-vif-plugged-b7e8fcf5-0b5e-4b35-af6b-f18163968d76 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.058852] env[62183]: DEBUG oslo_concurrency.lockutils [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] Acquiring lock "dbd34888-a333-40b4-ae9c-df5541b16704-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.059263] env[62183]: DEBUG oslo_concurrency.lockutils [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] Lock "dbd34888-a333-40b4-ae9c-df5541b16704-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.059481] env[62183]: DEBUG oslo_concurrency.lockutils [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] Lock "dbd34888-a333-40b4-ae9c-df5541b16704-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.059722] env[62183]: DEBUG nova.compute.manager [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] No waiting events found dispatching network-vif-plugged-b7e8fcf5-0b5e-4b35-af6b-f18163968d76 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1000.059855] env[62183]: WARNING nova.compute.manager [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Received unexpected event network-vif-plugged-b7e8fcf5-0b5e-4b35-af6b-f18163968d76 for instance with vm_state building and task_state spawning. [ 1000.060014] env[62183]: DEBUG nova.compute.manager [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Received event network-changed-b7e8fcf5-0b5e-4b35-af6b-f18163968d76 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.060195] env[62183]: DEBUG nova.compute.manager [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Refreshing instance network info cache due to event network-changed-b7e8fcf5-0b5e-4b35-af6b-f18163968d76. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1000.060371] env[62183]: DEBUG oslo_concurrency.lockutils [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] Acquiring lock "refresh_cache-dbd34888-a333-40b4-ae9c-df5541b16704" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.060509] env[62183]: DEBUG oslo_concurrency.lockutils [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] Acquired lock "refresh_cache-dbd34888-a333-40b4-ae9c-df5541b16704" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.060665] env[62183]: DEBUG nova.network.neutron [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Refreshing network info cache for port b7e8fcf5-0b5e-4b35-af6b-f18163968d76 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1000.082023] env[62183]: DEBUG oslo_concurrency.lockutils [req-86f4a87d-007c-4568-b205-db667845e5a9 req-37892996-5fbf-45c4-bdf5-c06f49ad15a4 service nova] Releasing lock "refresh_cache-a5018b4e-016d-4c61-bfbf-2c039ab79499" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.102899] env[62183]: DEBUG oslo_vmware.api [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387393, 'name': PowerOffVM_Task, 'duration_secs': 0.186273} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.103195] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.103373] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1000.103629] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6bb92b2-ce38-4ab6-89d3-2b0f90801016 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.141881] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525806a6-0eaf-4264-791d-3cbc426cc864, 'name': SearchDatastore_Task, 'duration_secs': 0.015031} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.142241] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.142494] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.142742] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.142915] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.143142] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.143416] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0930272-32fe-4eb1-ac13-37c8eaf69959 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.152539] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.152717] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1000.153434] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe79fc8e-50be-45aa-bca1-04008839d278 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.158507] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1000.158507] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5283612d-3760-5dcc-41a0-e5730c610c27" [ 1000.158507] env[62183]: _type = "Task" [ 1000.158507] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.165639] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5283612d-3760-5dcc-41a0-e5730c610c27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.179330] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1000.179545] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1000.179728] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleting the datastore file [datastore2] a948464b-63aa-4bc8-9885-228049e96d37 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1000.179991] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48f1d873-bc74-4cab-aa85-dbf3e803d803 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.185589] env[62183]: DEBUG oslo_vmware.api [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1000.185589] env[62183]: value = "task-1387396" [ 1000.185589] env[62183]: _type = "Task" [ 1000.185589] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.193458] env[62183]: DEBUG oslo_vmware.api [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387396, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.220015] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "refresh_cache-dbd34888-a333-40b4-ae9c-df5541b16704" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.257626] env[62183]: DEBUG nova.compute.utils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1000.262069] env[62183]: DEBUG nova.compute.manager [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1000.262283] env[62183]: DEBUG nova.network.neutron [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1000.303916] env[62183]: DEBUG nova.policy [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4f1ba64ecf948ddabab1c59745f9fc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '587870b99b844c819d27ac4852e44c3d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1000.334564] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387394, 'name': CreateVM_Task, 'duration_secs': 0.346939} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.334564] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1000.334859] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.335017] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.335348] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1000.335605] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d465473-0536-4ba5-aabc-83d3db5c342c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.343179] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1000.343179] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525bbb5e-e0b1-2dda-b31d-4225721df5f2" [ 1000.343179] env[62183]: _type = "Task" [ 1000.343179] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.351487] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525bbb5e-e0b1-2dda-b31d-4225721df5f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.483600] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6dc3cd5-fd79-461f-8d82-1e06024ca8f4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.492889] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba90081-a303-49d5-958c-21318fdafc47 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.520777] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a1a733-665c-4af0-bec6-e9b1b1331c89 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.528234] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b9d897-1139-4d1e-ac38-096c725b0b5c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.541405] env[62183]: DEBUG nova.compute.provider_tree [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.581356] env[62183]: DEBUG nova.network.neutron [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Successfully created port: 5034c986-efd5-4de5-8bc7-6391f23605cd {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1000.597874] env[62183]: DEBUG nova.network.neutron [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1000.669516] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5283612d-3760-5dcc-41a0-e5730c610c27, 'name': SearchDatastore_Task, 'duration_secs': 0.00839} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.670169] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cf50ffa-99b8-4d4c-b081-9bf301697808 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.672821] env[62183]: DEBUG nova.network.neutron [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.676756] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1000.676756] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52911ab4-13ee-db38-ff63-210db0644fce" [ 1000.676756] env[62183]: _type = "Task" [ 1000.676756] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.688255] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52911ab4-13ee-db38-ff63-210db0644fce, 'name': SearchDatastore_Task, 'duration_secs': 0.009058} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.691209] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.691466] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] a5018b4e-016d-4c61-bfbf-2c039ab79499/a5018b4e-016d-4c61-bfbf-2c039ab79499.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1000.691706] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-894dcbbb-3a22-4782-a5a7-56d4c24dc775 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.697864] env[62183]: DEBUG oslo_vmware.api [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387396, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13617} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.698982] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.699248] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.699438] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.699617] env[62183]: INFO nova.compute.manager [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1000.699846] env[62183]: DEBUG oslo.service.loopingcall [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.700158] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1000.700158] env[62183]: value = "task-1387397" [ 1000.700158] env[62183]: _type = "Task" [ 1000.700158] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.700423] env[62183]: DEBUG nova.compute.manager [-] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.700423] env[62183]: DEBUG nova.network.neutron [-] [instance: a948464b-63aa-4bc8-9885-228049e96d37] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1000.709380] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387397, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.762997] env[62183]: DEBUG nova.compute.manager [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1000.853780] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525bbb5e-e0b1-2dda-b31d-4225721df5f2, 'name': SearchDatastore_Task, 'duration_secs': 0.008953} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.854127] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.854374] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.854636] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.854803] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.854995] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.855355] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45061b75-ef92-4952-86d5-7a727ca57715 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.868720] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.868944] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1000.869780] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-628c9824-2d57-4ee0-b8ef-966df13ad5eb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.875358] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1000.875358] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5255bf9e-8934-f582-567a-42e03b8142b3" [ 1000.875358] env[62183]: _type = "Task" [ 1000.875358] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.883231] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5255bf9e-8934-f582-567a-42e03b8142b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.046525] env[62183]: DEBUG nova.scheduler.client.report [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1001.175629] env[62183]: DEBUG oslo_concurrency.lockutils [req-de02cfa6-ffc5-43a6-ae05-457a3aaba1d9 req-b2f73c36-acd8-4d00-87a9-6c90c63f9962 service nova] Releasing lock "refresh_cache-dbd34888-a333-40b4-ae9c-df5541b16704" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.176049] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "refresh_cache-dbd34888-a333-40b4-ae9c-df5541b16704" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.176225] env[62183]: DEBUG nova.network.neutron [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.193113] env[62183]: DEBUG nova.compute.manager [req-d66f3347-7446-4d9d-9f78-6c577e8c303b req-9d084e23-216a-429f-ac85-501697d8b0cb service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Received event network-vif-deleted-3d687443-41fd-44ff-8981-35319b79d88a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.193225] env[62183]: INFO nova.compute.manager [req-d66f3347-7446-4d9d-9f78-6c577e8c303b req-9d084e23-216a-429f-ac85-501697d8b0cb service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Neutron deleted interface 3d687443-41fd-44ff-8981-35319b79d88a; detaching it from the instance and deleting it from the info cache [ 1001.193365] env[62183]: DEBUG nova.network.neutron [req-d66f3347-7446-4d9d-9f78-6c577e8c303b req-9d084e23-216a-429f-ac85-501697d8b0cb service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.213918] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387397, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477754} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.214169] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] a5018b4e-016d-4c61-bfbf-2c039ab79499/a5018b4e-016d-4c61-bfbf-2c039ab79499.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1001.214406] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1001.214659] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70c715e5-4bbe-4d66-baa5-aa24a584fa18 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.221310] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1001.221310] env[62183]: value = "task-1387398" [ 1001.221310] env[62183]: _type = "Task" [ 1001.221310] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.229107] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387398, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.389301] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5255bf9e-8934-f582-567a-42e03b8142b3, 'name': SearchDatastore_Task, 'duration_secs': 0.057165} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.389301] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbb1a02d-20d8-4575-84a0-9c0417958479 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.394024] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1001.394024] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]526ddcf9-480d-c395-442c-01d82fce50b1" [ 1001.394024] env[62183]: _type = "Task" [ 1001.394024] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.401140] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526ddcf9-480d-c395-442c-01d82fce50b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.551713] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.798s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.554094] env[62183]: DEBUG oslo_concurrency.lockutils [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.168s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.554333] env[62183]: DEBUG nova.objects.instance [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lazy-loading 'resources' on Instance uuid 143323db-da1c-4dd8-8f13-ed0af31e2027 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.573258] env[62183]: INFO nova.scheduler.client.report [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleted allocations for instance b6402e2f-1035-4c1b-9a1e-05b17d89e4bb [ 1001.671312] env[62183]: DEBUG nova.network.neutron [-] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.697238] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03bfb8c2-4db4-4b77-9b3e-6ec6158ca2e9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.706575] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90d3b1d-bc20-4215-a078-9e5fc8e0f147 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.718034] env[62183]: DEBUG nova.network.neutron [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1001.729736] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387398, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064336} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.737883] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.738296] env[62183]: DEBUG nova.compute.manager [req-d66f3347-7446-4d9d-9f78-6c577e8c303b req-9d084e23-216a-429f-ac85-501697d8b0cb service nova] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Detach interface failed, port_id=3d687443-41fd-44ff-8981-35319b79d88a, reason: Instance a948464b-63aa-4bc8-9885-228049e96d37 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1001.741121] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c8a4d8-0331-4638-b13f-bdc4a2cacd86 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.763167] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] a5018b4e-016d-4c61-bfbf-2c039ab79499/a5018b4e-016d-4c61-bfbf-2c039ab79499.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.765491] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9342a1c5-6d01-4a52-9842-62b8cfcb139e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.779881] env[62183]: DEBUG nova.compute.manager [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1001.786982] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1001.786982] env[62183]: value = "task-1387399" [ 1001.786982] env[62183]: _type = "Task" [ 1001.786982] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.794801] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387399, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.803725] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.803954] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.804163] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.804330] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.804479] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.804687] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.804906] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.805099] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.805281] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.805449] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.805624] env[62183]: DEBUG nova.virt.hardware [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.806434] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ec83ae-241f-4572-9f53-5ecbc2368a46 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.817407] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f18cbd-ac62-482f-98d2-3be27bba14e5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.882211] env[62183]: DEBUG nova.network.neutron [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Updating instance_info_cache with network_info: [{"id": "b7e8fcf5-0b5e-4b35-af6b-f18163968d76", "address": "fa:16:3e:23:84:21", "network": {"id": "7cbe5832-931e-430d-95f5-8f48781f095b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-236543802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "587870b99b844c819d27ac4852e44c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7e8fcf5-0b", "ovs_interfaceid": "b7e8fcf5-0b5e-4b35-af6b-f18163968d76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.905189] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]526ddcf9-480d-c395-442c-01d82fce50b1, 'name': SearchDatastore_Task, 'duration_secs': 0.009189} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.905447] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.905708] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4/d8817d24-b0cb-4956-b195-cc417ae09fb4.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1001.905961] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca86331f-9328-46c1-acaf-142824ccb72e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.912239] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1001.912239] env[62183]: value = "task-1387400" [ 1001.912239] env[62183]: _type = "Task" [ 1001.912239] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.920639] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387400, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.056048] env[62183]: DEBUG nova.network.neutron [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Successfully updated port: 5034c986-efd5-4de5-8bc7-6391f23605cd {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1002.080448] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a63f273a-e1d8-4e46-be22-114d46261530 tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "b6402e2f-1035-4c1b-9a1e-05b17d89e4bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.704s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.109123] env[62183]: DEBUG nova.compute.manager [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Received event network-vif-plugged-5034c986-efd5-4de5-8bc7-6391f23605cd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.109432] env[62183]: DEBUG oslo_concurrency.lockutils [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] Acquiring lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.109581] env[62183]: DEBUG oslo_concurrency.lockutils [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] Lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.109743] env[62183]: DEBUG oslo_concurrency.lockutils [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] Lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.111452] env[62183]: DEBUG nova.compute.manager [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] No waiting events found dispatching network-vif-plugged-5034c986-efd5-4de5-8bc7-6391f23605cd {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1002.111687] env[62183]: WARNING nova.compute.manager [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Received unexpected event network-vif-plugged-5034c986-efd5-4de5-8bc7-6391f23605cd for instance with vm_state building and task_state spawning. [ 1002.111878] env[62183]: DEBUG nova.compute.manager [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Received event network-changed-5034c986-efd5-4de5-8bc7-6391f23605cd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.112064] env[62183]: DEBUG nova.compute.manager [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Refreshing instance network info cache due to event network-changed-5034c986-efd5-4de5-8bc7-6391f23605cd. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.112269] env[62183]: DEBUG oslo_concurrency.lockutils [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] Acquiring lock "refresh_cache-ac58afb9-9f4c-4092-bcfc-b3ed1433a356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.112435] env[62183]: DEBUG oslo_concurrency.lockutils [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] Acquired lock "refresh_cache-ac58afb9-9f4c-4092-bcfc-b3ed1433a356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.112608] env[62183]: DEBUG nova.network.neutron [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Refreshing network info cache for port 5034c986-efd5-4de5-8bc7-6391f23605cd {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.175142] env[62183]: INFO nova.compute.manager [-] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Took 1.47 seconds to deallocate network for instance. [ 1002.267446] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e84ee3-281f-4213-9944-1982c939682f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.275385] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4059b9d-1ccd-41c0-aa5c-3b8463570c3c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.312381] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380f174a-6bde-4731-bcd9-c667f7884553 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.320510] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387399, 'name': ReconfigVM_Task, 'duration_secs': 0.263869} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.322926] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Reconfigured VM instance instance-00000062 to attach disk [datastore1] a5018b4e-016d-4c61-bfbf-2c039ab79499/a5018b4e-016d-4c61-bfbf-2c039ab79499.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.323708] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95913efa-fa06-4cc8-b790-fe7915ec114f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.326492] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f6b8cc-cf29-4b99-afc5-ea55a6cdf407 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.344367] env[62183]: DEBUG nova.compute.provider_tree [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.347390] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1002.347390] env[62183]: value = "task-1387401" [ 1002.347390] env[62183]: _type = "Task" [ 1002.347390] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.356389] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387401, 'name': Rename_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.385409] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "refresh_cache-dbd34888-a333-40b4-ae9c-df5541b16704" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.385786] env[62183]: DEBUG nova.compute.manager [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Instance network_info: |[{"id": "b7e8fcf5-0b5e-4b35-af6b-f18163968d76", "address": "fa:16:3e:23:84:21", "network": {"id": "7cbe5832-931e-430d-95f5-8f48781f095b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-236543802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "587870b99b844c819d27ac4852e44c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7e8fcf5-0b", "ovs_interfaceid": "b7e8fcf5-0b5e-4b35-af6b-f18163968d76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1002.386336] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:84:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e55c248-c504-4c7a-bbe9-f42cf417aee7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7e8fcf5-0b5e-4b35-af6b-f18163968d76', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.394348] env[62183]: DEBUG oslo.service.loopingcall [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.394954] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1002.395203] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c85046e7-1196-429b-815e-77fe50397663 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.417742] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.417742] env[62183]: value = "task-1387402" [ 1002.417742] env[62183]: _type = "Task" [ 1002.417742] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.424857] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387400, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.430583] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387402, 'name': CreateVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.561682] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "refresh_cache-ac58afb9-9f4c-4092-bcfc-b3ed1433a356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.648295] env[62183]: DEBUG nova.network.neutron [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1002.683550] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.734573] env[62183]: DEBUG nova.network.neutron [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.849532] env[62183]: DEBUG nova.scheduler.client.report [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1002.862856] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387401, 'name': Rename_Task, 'duration_secs': 0.237339} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.863217] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1002.863485] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc1412a4-3c54-4228-8e8b-4c57f285702e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.870718] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1002.870718] env[62183]: value = "task-1387403" [ 1002.870718] env[62183]: _type = "Task" [ 1002.870718] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.879798] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387403, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.925793] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387400, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589337} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.926114] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4/d8817d24-b0cb-4956-b195-cc417ae09fb4.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1002.926341] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1002.927060] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28910ecb-5402-484b-9683-d360336d3529 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.932239] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387402, 'name': CreateVM_Task, 'duration_secs': 0.505637} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.932711] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1002.933419] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.933581] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.934292] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.934292] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae657c35-7795-47f3-91e3-820180c6b321 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.938470] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1002.938470] env[62183]: value = "task-1387404" [ 1002.938470] env[62183]: _type = "Task" [ 1002.938470] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.943988] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1002.943988] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52a33152-3501-d5b3-7ff2-d29ace54ea36" [ 1002.943988] env[62183]: _type = "Task" [ 1002.943988] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.951770] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387404, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.956818] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a33152-3501-d5b3-7ff2-d29ace54ea36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.237984] env[62183]: DEBUG oslo_concurrency.lockutils [req-98f1ed53-22e3-4947-bb45-03ba8f853fe6 req-5b9b25c4-8306-4579-a36f-d4c0163c5ac5 service nova] Releasing lock "refresh_cache-ac58afb9-9f4c-4092-bcfc-b3ed1433a356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.238477] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "refresh_cache-ac58afb9-9f4c-4092-bcfc-b3ed1433a356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.238655] env[62183]: DEBUG nova.network.neutron [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1003.358982] env[62183]: DEBUG oslo_concurrency.lockutils [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.804s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.361884] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.018s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.362151] env[62183]: DEBUG nova.objects.instance [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'resources' on Instance uuid c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.380660] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387403, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.390710] env[62183]: INFO nova.scheduler.client.report [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Deleted allocations for instance 143323db-da1c-4dd8-8f13-ed0af31e2027 [ 1003.453159] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387404, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.136363} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.457105] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.457424] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52a33152-3501-d5b3-7ff2-d29ace54ea36, 'name': SearchDatastore_Task, 'duration_secs': 0.014112} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.458446] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb6fb6f-b2be-4a08-96c0-19e057a43686 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.460503] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.460733] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.460965] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.461133] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.461316] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.461560] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-beb729f7-181f-4066-b044-6f82b1594df2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.482798] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4/d8817d24-b0cb-4956-b195-cc417ae09fb4.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.484236] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84f29628-1178-432c-b860-9acdf696172e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.498324] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.498528] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.499283] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df66a21b-c4d5-45f2-9150-de2bedfdbe2f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.504972] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1003.504972] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]521c3df8-37ab-cf04-4783-c930b7a375f5" [ 1003.504972] env[62183]: _type = "Task" [ 1003.504972] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.506646] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1003.506646] env[62183]: value = "task-1387405" [ 1003.506646] env[62183]: _type = "Task" [ 1003.506646] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.517728] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521c3df8-37ab-cf04-4783-c930b7a375f5, 'name': SearchDatastore_Task, 'duration_secs': 0.008327} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.521362] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387405, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.521587] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-613fd627-f9c6-4742-a2b5-13af32dcaf4a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.526450] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1003.526450] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52b3f37f-7387-43b9-d5a9-f113ca1bef19" [ 1003.526450] env[62183]: _type = "Task" [ 1003.526450] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.533872] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b3f37f-7387-43b9-d5a9-f113ca1bef19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.772337] env[62183]: DEBUG nova.network.neutron [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1003.879832] env[62183]: DEBUG oslo_vmware.api [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387403, 'name': PowerOnVM_Task, 'duration_secs': 0.808719} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.880128] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1003.880334] env[62183]: INFO nova.compute.manager [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1003.880595] env[62183]: DEBUG nova.compute.manager [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.881287] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f9484e-e84c-432c-a11f-d10111b84430 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.899829] env[62183]: DEBUG nova.network.neutron [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Updating instance_info_cache with network_info: [{"id": "5034c986-efd5-4de5-8bc7-6391f23605cd", "address": "fa:16:3e:c9:bc:6d", "network": {"id": "7cbe5832-931e-430d-95f5-8f48781f095b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-236543802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "587870b99b844c819d27ac4852e44c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5034c986-ef", "ovs_interfaceid": "5034c986-efd5-4de5-8bc7-6391f23605cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.900985] env[62183]: DEBUG oslo_concurrency.lockutils [None req-62f3cdf9-0ff2-4792-a8ad-2bcac7e8248a tempest-MultipleCreateTestJSON-782368991 tempest-MultipleCreateTestJSON-782368991-project-member] Lock "143323db-da1c-4dd8-8f13-ed0af31e2027" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.455s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.021514] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387405, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.035687] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52b3f37f-7387-43b9-d5a9-f113ca1bef19, 'name': SearchDatastore_Task, 'duration_secs': 0.008299} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.036812] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.037098] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] dbd34888-a333-40b4-ae9c-df5541b16704/dbd34888-a333-40b4-ae9c-df5541b16704.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.037916] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0409438-ad22-4b39-a821-435f60bcb130 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.040423] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2cbfd45b-4d08-4e92-948d-b1031786a319 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.046747] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f594b9-b56a-4ed3-a737-5ef0be18b1a3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.050704] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1004.050704] env[62183]: value = "task-1387406" [ 1004.050704] env[62183]: _type = "Task" [ 1004.050704] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.080431] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41749790-6f9e-4f99-8cdc-2a1c28e6a006 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.085957] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387406, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.091197] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473e6aee-ecaa-471f-941e-93efd4acc5b3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.104232] env[62183]: DEBUG nova.compute.provider_tree [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.401030] env[62183]: INFO nova.compute.manager [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Took 13.66 seconds to build instance. [ 1004.405459] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "refresh_cache-ac58afb9-9f4c-4092-bcfc-b3ed1433a356" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.405459] env[62183]: DEBUG nova.compute.manager [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Instance network_info: |[{"id": "5034c986-efd5-4de5-8bc7-6391f23605cd", "address": "fa:16:3e:c9:bc:6d", "network": {"id": "7cbe5832-931e-430d-95f5-8f48781f095b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-236543802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "587870b99b844c819d27ac4852e44c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5034c986-ef", "ovs_interfaceid": "5034c986-efd5-4de5-8bc7-6391f23605cd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1004.405459] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:bc:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e55c248-c504-4c7a-bbe9-f42cf417aee7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5034c986-efd5-4de5-8bc7-6391f23605cd', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1004.413419] env[62183]: DEBUG oslo.service.loopingcall [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.413616] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1004.414560] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0911679-f27b-4091-b6fc-32c5290e6909 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.436659] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1004.436659] env[62183]: value = "task-1387407" [ 1004.436659] env[62183]: _type = "Task" [ 1004.436659] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.446860] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387407, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.521328] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387405, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.560253] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387406, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463165} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.560497] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] dbd34888-a333-40b4-ae9c-df5541b16704/dbd34888-a333-40b4-ae9c-df5541b16704.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1004.560717] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.560984] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30ecadcd-bb72-47e0-bb5a-662aa1f62ebd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.567426] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1004.567426] env[62183]: value = "task-1387408" [ 1004.567426] env[62183]: _type = "Task" [ 1004.567426] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.575017] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387408, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.607404] env[62183]: DEBUG nova.scheduler.client.report [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1004.720561] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "8c4b68d6-fedd-408f-a449-aace7400014a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.720763] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.904577] env[62183]: DEBUG oslo_concurrency.lockutils [None req-962c9432-870c-4e72-a9bc-72562afb7087 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.176s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.946937] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387407, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.021948] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387405, 'name': ReconfigVM_Task, 'duration_secs': 1.309947} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.022205] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Reconfigured VM instance instance-00000061 to attach disk [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4/d8817d24-b0cb-4956-b195-cc417ae09fb4.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.022829] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d4aebb2-b7d3-4602-a729-f9760e6ed6f0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.028478] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1005.028478] env[62183]: value = "task-1387409" [ 1005.028478] env[62183]: _type = "Task" [ 1005.028478] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.035731] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387409, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.076238] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387408, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061154} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.076515] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.077328] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98debccd-8b0d-4eee-ac80-3b499aef9f51 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.098585] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] dbd34888-a333-40b4-ae9c-df5541b16704/dbd34888-a333-40b4-ae9c-df5541b16704.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.098885] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4d785fd-606f-4780-9124-915f25e11b61 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.112948] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.751s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.115129] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.514s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.115372] env[62183]: DEBUG nova.objects.instance [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lazy-loading 'resources' on Instance uuid 036037b3-9676-419c-ab87-e5a557438fbe {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.121626] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1005.121626] env[62183]: value = "task-1387410" [ 1005.121626] env[62183]: _type = "Task" [ 1005.121626] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.130992] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387410, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.136858] env[62183]: INFO nova.scheduler.client.report [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleted allocations for instance c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d [ 1005.226485] env[62183]: DEBUG nova.compute.utils [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.447912] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387407, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.538177] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387409, 'name': Rename_Task, 'duration_secs': 0.150317} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.538479] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.538759] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1322fc7d-0c11-4bca-ba32-9959aee2dea9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.545626] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1005.545626] env[62183]: value = "task-1387411" [ 1005.545626] env[62183]: _type = "Task" [ 1005.545626] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.555150] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387411, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.632614] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387410, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.647889] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a77605e0-368d-40c5-8d56-30898b523235 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.486s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.729358] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.877392] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef42d3c8-0313-4164-922b-fa170efc7905 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.886447] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a8b5cb-8040-42f0-a70b-d45fae1bc7c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.918108] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6af899-6644-4800-9c01-c5f6dc620dcc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.926027] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75293556-db80-4f6a-b671-791e671471cc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.941840] env[62183]: DEBUG nova.compute.provider_tree [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.954467] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387407, 'name': CreateVM_Task, 'duration_secs': 1.096716} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.955096] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1005.955870] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.956145] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.956582] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1005.957194] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07278728-bb66-4ff4-920b-746a3f96e68e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.961900] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1005.961900] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52553bf6-e69b-3f5c-64e6-56bf2e80cf7e" [ 1005.961900] env[62183]: _type = "Task" [ 1005.961900] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.972658] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52553bf6-e69b-3f5c-64e6-56bf2e80cf7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.056235] env[62183]: DEBUG oslo_vmware.api [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387411, 'name': PowerOnVM_Task, 'duration_secs': 0.430638} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.056866] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.056866] env[62183]: DEBUG nova.compute.manager [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.057661] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd448e5-e2e5-4cc6-a1ad-e2e05ed9a20c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.131745] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387410, 'name': ReconfigVM_Task, 'duration_secs': 0.900359} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.132110] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Reconfigured VM instance instance-00000063 to attach disk [datastore1] dbd34888-a333-40b4-ae9c-df5541b16704/dbd34888-a333-40b4-ae9c-df5541b16704.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.132738] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-daf7f7b9-7af8-42dd-b936-7fe0e859b76c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.143091] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1006.143091] env[62183]: value = "task-1387412" [ 1006.143091] env[62183]: _type = "Task" [ 1006.143091] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.154481] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387412, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.448166] env[62183]: DEBUG nova.scheduler.client.report [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1006.473786] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52553bf6-e69b-3f5c-64e6-56bf2e80cf7e, 'name': SearchDatastore_Task, 'duration_secs': 0.008947} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.474152] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.474409] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.474647] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.474796] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.474975] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.475283] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40b2cad0-803d-4501-881b-ecc8bbf50d25 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.484356] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.484544] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1006.485301] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f363a92-2886-40a0-87cf-f0fee05b28d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.491092] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1006.491092] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52c9a550-5662-cc11-13d4-b0058fa15a29" [ 1006.491092] env[62183]: _type = "Task" [ 1006.491092] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.498556] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c9a550-5662-cc11-13d4-b0058fa15a29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.575979] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.652679] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387412, 'name': Rename_Task, 'duration_secs': 0.151487} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.652955] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1006.653219] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13adcb45-008a-4111-be56-36f21c9c0f72 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.659295] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1006.659295] env[62183]: value = "task-1387413" [ 1006.659295] env[62183]: _type = "Task" [ 1006.659295] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.669068] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387413, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.737019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.737019] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.803022] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "8c4b68d6-fedd-408f-a449-aace7400014a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.803022] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.803022] env[62183]: INFO nova.compute.manager [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Attaching volume 786b9d04-6ed3-4cb1-a773-6db867ff2adb to /dev/sdb [ 1006.841505] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93400a4b-ec37-4020-98d8-ba8739d72bb0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.848826] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbc127b-ac67-4c4b-9ff8-bab71c6742dd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.862926] env[62183]: DEBUG nova.virt.block_device [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updating existing volume attachment record: 6e403644-624e-4cae-b2cc-33bd1ffd92ee {{(pid=62183) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1006.953966] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.839s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.956355] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.208s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.956598] env[62183]: DEBUG nova.objects.instance [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lazy-loading 'resources' on Instance uuid 24f7c072-1209-4c6e-9a54-4e069f14f7d2 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.985406] env[62183]: INFO nova.scheduler.client.report [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Deleted allocations for instance 036037b3-9676-419c-ab87-e5a557438fbe [ 1007.000976] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c9a550-5662-cc11-13d4-b0058fa15a29, 'name': SearchDatastore_Task, 'duration_secs': 0.00823} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.002186] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5a58ba6-129d-4687-ad7f-e69a3824e710 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.007462] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1007.007462] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5233723b-02dc-db80-a944-f038694efc01" [ 1007.007462] env[62183]: _type = "Task" [ 1007.007462] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.017221] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5233723b-02dc-db80-a944-f038694efc01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.170367] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387413, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.238735] env[62183]: INFO nova.compute.manager [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Detaching volume f481c3e9-7f36-4566-9973-b98f2c584ee9 [ 1007.285021] env[62183]: INFO nova.virt.block_device [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Attempting to driver detach volume f481c3e9-7f36-4566-9973-b98f2c584ee9 from mountpoint /dev/sdb [ 1007.285021] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Volume detach. Driver type: vmdk {{(pid=62183) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1007.285021] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294482', 'volume_id': 'f481c3e9-7f36-4566-9973-b98f2c584ee9', 'name': 'volume-f481c3e9-7f36-4566-9973-b98f2c584ee9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de', 'attached_at': '', 'detached_at': '', 'volume_id': 'f481c3e9-7f36-4566-9973-b98f2c584ee9', 'serial': 'f481c3e9-7f36-4566-9973-b98f2c584ee9'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1007.285921] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c707fee-f2b7-4da3-ac0e-48a265076c02 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.313387] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a7ac2b-24f9-4460-bd8b-8588d01a4392 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.326056] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-860152ae-e287-4f59-9433-97f0220a9a88 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.345259] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb25bea-911b-47a1-86b3-8280cf42c090 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.361212] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] The volume has not been displaced from its original location: [datastore1] volume-f481c3e9-7f36-4566-9973-b98f2c584ee9/volume-f481c3e9-7f36-4566-9973-b98f2c584ee9.vmdk. No consolidation needed. {{(pid=62183) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1007.369342] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Reconfiguring VM instance instance-0000004f to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1007.369342] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25dadb02-4a25-47fe-ab22-9333b64b98a8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.389970] env[62183]: DEBUG oslo_vmware.api [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1007.389970] env[62183]: value = "task-1387417" [ 1007.389970] env[62183]: _type = "Task" [ 1007.389970] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.400754] env[62183]: DEBUG oslo_vmware.api [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387417, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.496991] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0f39829d-062c-4717-b60d-0c14865db689 tempest-ServerAddressesTestJSON-1304786800 tempest-ServerAddressesTestJSON-1304786800-project-member] Lock "036037b3-9676-419c-ab87-e5a557438fbe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.903s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.528186] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5233723b-02dc-db80-a944-f038694efc01, 'name': SearchDatastore_Task, 'duration_secs': 0.009305} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.529950] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.530616] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] ac58afb9-9f4c-4092-bcfc-b3ed1433a356/ac58afb9-9f4c-4092-bcfc-b3ed1433a356.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1007.535205] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3018e269-94a7-4f9e-aebd-54027c5ba106 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.545995] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1007.545995] env[62183]: value = "task-1387418" [ 1007.545995] env[62183]: _type = "Task" [ 1007.545995] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.558167] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.674065] env[62183]: DEBUG oslo_vmware.api [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387413, 'name': PowerOnVM_Task, 'duration_secs': 0.754104} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.674552] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1007.675063] env[62183]: INFO nova.compute.manager [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Took 8.23 seconds to spawn the instance on the hypervisor. [ 1007.675389] env[62183]: DEBUG nova.compute.manager [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.679749] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a59474-050c-4d9b-be78-4834d35c7325 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.698043] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda501fc-215f-43b2-bbda-79d85e5f8d7a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.710044] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba86759-86c5-45b9-b422-9ecb1df5dabf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.748251] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "d8817d24-b0cb-4956-b195-cc417ae09fb4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.748627] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "d8817d24-b0cb-4956-b195-cc417ae09fb4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.748911] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "d8817d24-b0cb-4956-b195-cc417ae09fb4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.749292] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "d8817d24-b0cb-4956-b195-cc417ae09fb4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.749568] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "d8817d24-b0cb-4956-b195-cc417ae09fb4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.752687] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0fa80e2-3962-42d5-8910-a5870528820d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.755545] env[62183]: INFO nova.compute.manager [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Terminating instance [ 1007.757676] env[62183]: DEBUG nova.compute.manager [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1007.757931] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.759105] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c8112c-e33a-4dab-885e-67de5f2d1bd0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.765640] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac1dab7-0414-4189-917d-229e8409de96 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.771690] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.772623] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e57ea68-2e45-4ffa-8d07-b110492cae5b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.782312] env[62183]: DEBUG nova.compute.provider_tree [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.787019] env[62183]: DEBUG oslo_vmware.api [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1007.787019] env[62183]: value = "task-1387419" [ 1007.787019] env[62183]: _type = "Task" [ 1007.787019] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.794546] env[62183]: DEBUG oslo_vmware.api [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387419, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.900392] env[62183]: DEBUG oslo_vmware.api [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387417, 'name': ReconfigVM_Task, 'duration_secs': 0.255798} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.900798] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Reconfigured VM instance instance-0000004f to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1007.905524] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1c1294d-f3c5-45db-9ba9-3b4765e74f16 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.923731] env[62183]: DEBUG oslo_vmware.api [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1007.923731] env[62183]: value = "task-1387420" [ 1007.923731] env[62183]: _type = "Task" [ 1007.923731] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.933338] env[62183]: DEBUG oslo_vmware.api [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387420, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.060022] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.199976] env[62183]: INFO nova.compute.manager [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Took 17.14 seconds to build instance. [ 1008.287017] env[62183]: DEBUG nova.scheduler.client.report [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1008.300515] env[62183]: DEBUG oslo_vmware.api [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387419, 'name': PowerOffVM_Task, 'duration_secs': 0.192346} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.300810] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1008.300987] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.301300] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d11854c-0f41-493d-93f2-9a55a14f3312 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.368643] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.368851] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.368991] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleting the datastore file [datastore1] d8817d24-b0cb-4956-b195-cc417ae09fb4 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.369980] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98cf3a68-f2cb-4578-9868-a71cdd92b79f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.378962] env[62183]: DEBUG oslo_vmware.api [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1008.378962] env[62183]: value = "task-1387422" [ 1008.378962] env[62183]: _type = "Task" [ 1008.378962] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.386384] env[62183]: DEBUG oslo_vmware.api [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.434418] env[62183]: DEBUG oslo_vmware.api [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387420, 'name': ReconfigVM_Task, 'duration_secs': 0.179814} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.434418] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294482', 'volume_id': 'f481c3e9-7f36-4566-9973-b98f2c584ee9', 'name': 'volume-f481c3e9-7f36-4566-9973-b98f2c584ee9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de', 'attached_at': '', 'detached_at': '', 'volume_id': 'f481c3e9-7f36-4566-9973-b98f2c584ee9', 'serial': 'f481c3e9-7f36-4566-9973-b98f2c584ee9'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1008.559257] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387418, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.701669] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1207376-3875-401d-8ec4-eb86401dc686 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "dbd34888-a333-40b4-ae9c-df5541b16704" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.645s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.795406] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.839s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.801029] env[62183]: DEBUG oslo_concurrency.lockutils [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.460s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.801029] env[62183]: DEBUG nova.objects.instance [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lazy-loading 'resources' on Instance uuid fb348784-62a0-4d1f-ac7f-f176f3da0dd9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.825385] env[62183]: INFO nova.scheduler.client.report [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted allocations for instance 24f7c072-1209-4c6e-9a54-4e069f14f7d2 [ 1008.892625] env[62183]: DEBUG oslo_vmware.api [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186105} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.894766] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.894766] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.894766] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.894766] env[62183]: INFO nova.compute.manager [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1008.894766] env[62183]: DEBUG oslo.service.loopingcall [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.894766] env[62183]: DEBUG nova.compute.manager [-] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1008.894766] env[62183]: DEBUG nova.network.neutron [-] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1008.993986] env[62183]: DEBUG nova.objects.instance [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'flavor' on Instance uuid ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.058288] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387418, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.371927} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.058786] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] ac58afb9-9f4c-4092-bcfc-b3ed1433a356/ac58afb9-9f4c-4092-bcfc-b3ed1433a356.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1009.059175] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1009.059564] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78990956-b327-4cf3-aa2c-ad294496192a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.066081] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1009.066081] env[62183]: value = "task-1387423" [ 1009.066081] env[62183]: _type = "Task" [ 1009.066081] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.080199] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387423, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.297128] env[62183]: DEBUG nova.compute.manager [req-91b3cee1-1239-4118-ba16-31e5405ea047 req-0c906fc5-ca37-4b6b-a37a-dfb91d5dde9a service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Received event network-vif-deleted-8b2e8106-ed68-4023-b3cc-d91842d269be {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.297128] env[62183]: INFO nova.compute.manager [req-91b3cee1-1239-4118-ba16-31e5405ea047 req-0c906fc5-ca37-4b6b-a37a-dfb91d5dde9a service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Neutron deleted interface 8b2e8106-ed68-4023-b3cc-d91842d269be; detaching it from the instance and deleting it from the info cache [ 1009.297128] env[62183]: DEBUG nova.network.neutron [req-91b3cee1-1239-4118-ba16-31e5405ea047 req-0c906fc5-ca37-4b6b-a37a-dfb91d5dde9a service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.334471] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0225544f-094e-4320-90a7-f0c87292c7ae tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "24f7c072-1209-4c6e-9a54-4e069f14f7d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.616s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.580992] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387423, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068964} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.580992] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1009.580992] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bb5cc2-4d80-4a96-a458-e7322847313d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.608163] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] ac58afb9-9f4c-4092-bcfc-b3ed1433a356/ac58afb9-9f4c-4092-bcfc-b3ed1433a356.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1009.613021] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a2f0cfe-7071-41c3-8b1d-c9e6da97dbfe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.629016] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990608c8-5400-4467-a215-acd0d8693784 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.637685] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4381e38-609f-43bc-b4ec-acf1ef82c71e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.640931] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1009.640931] env[62183]: value = "task-1387425" [ 1009.640931] env[62183]: _type = "Task" [ 1009.640931] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.674038] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b130a4b9-d653-401d-b58c-13ec9f1715af {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.677688] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387425, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.682438] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4453c7b4-260a-4e04-937f-fa62d17b44f6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.699383] env[62183]: DEBUG nova.compute.provider_tree [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.764819] env[62183]: DEBUG nova.network.neutron [-] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.806160] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-99bdbf5d-3374-4e6f-a8d4-d6d7fb4a8a28 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.819489] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e030c8a5-fd55-40ad-8b2b-4e52cbfd553c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.854362] env[62183]: DEBUG nova.compute.manager [req-91b3cee1-1239-4118-ba16-31e5405ea047 req-0c906fc5-ca37-4b6b-a37a-dfb91d5dde9a service nova] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Detach interface failed, port_id=8b2e8106-ed68-4023-b3cc-d91842d269be, reason: Instance d8817d24-b0cb-4956-b195-cc417ae09fb4 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1010.008123] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fbbf1f55-3eed-48d5-96f3-a20bfe0b72ba tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.273s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.155724] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387425, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.204236] env[62183]: DEBUG nova.scheduler.client.report [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1010.270681] env[62183]: INFO nova.compute.manager [-] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Took 1.38 seconds to deallocate network for instance. [ 1010.288782] env[62183]: DEBUG oslo_concurrency.lockutils [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "3a3ca703-2883-4aa9-a33f-326cc25d4838" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.289275] env[62183]: DEBUG oslo_concurrency.lockutils [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "3a3ca703-2883-4aa9-a33f-326cc25d4838" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.289772] env[62183]: DEBUG oslo_concurrency.lockutils [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "3a3ca703-2883-4aa9-a33f-326cc25d4838-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.290283] env[62183]: DEBUG oslo_concurrency.lockutils [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "3a3ca703-2883-4aa9-a33f-326cc25d4838-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.291504] env[62183]: DEBUG oslo_concurrency.lockutils [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "3a3ca703-2883-4aa9-a33f-326cc25d4838-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.295128] env[62183]: INFO nova.compute.manager [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Terminating instance [ 1010.299151] env[62183]: DEBUG nova.compute.manager [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.299462] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1010.300778] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e762fec8-9ed4-4b32-a3ac-8bcc0bfb144f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.310368] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1010.310368] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cd8db235-100f-4f48-b1b1-1e95ff34a797 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.318125] env[62183]: DEBUG oslo_vmware.api [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1010.318125] env[62183]: value = "task-1387426" [ 1010.318125] env[62183]: _type = "Task" [ 1010.318125] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.327596] env[62183]: DEBUG oslo_vmware.api [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387426, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.653705] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387425, 'name': ReconfigVM_Task, 'duration_secs': 0.543637} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.654042] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Reconfigured VM instance instance-00000064 to attach disk [datastore1] ac58afb9-9f4c-4092-bcfc-b3ed1433a356/ac58afb9-9f4c-4092-bcfc-b3ed1433a356.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1010.655585] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3c7f80d-a8fd-4f03-b227-a19dd02ba4f2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.662036] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1010.662036] env[62183]: value = "task-1387427" [ 1010.662036] env[62183]: _type = "Task" [ 1010.662036] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.670807] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387427, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.711929] env[62183]: DEBUG oslo_concurrency.lockutils [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.914s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.717845] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.034s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.718189] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.720161] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.144s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.720540] env[62183]: DEBUG nova.objects.instance [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62183) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1010.753070] env[62183]: INFO nova.scheduler.client.report [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Deleted allocations for instance fb348784-62a0-4d1f-ac7f-f176f3da0dd9 [ 1010.757858] env[62183]: INFO nova.scheduler.client.report [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted allocations for instance a948464b-63aa-4bc8-9885-228049e96d37 [ 1010.779800] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.828640] env[62183]: DEBUG oslo_vmware.api [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387426, 'name': PowerOffVM_Task, 'duration_secs': 0.165236} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.828849] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.828920] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.829222] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8194d58-221e-45be-bb70-dfb83238ca8b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.892284] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.892284] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.892284] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleting the datastore file [datastore1] 3a3ca703-2883-4aa9-a33f-326cc25d4838 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.892284] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46bb6377-5cb6-4e70-baa7-ea05c622ab83 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.899038] env[62183]: DEBUG oslo_vmware.api [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1010.899038] env[62183]: value = "task-1387429" [ 1010.899038] env[62183]: _type = "Task" [ 1010.899038] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.906748] env[62183]: DEBUG oslo_vmware.api [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387429, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.146874] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.147043] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.148084] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.148084] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.148084] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.150923] env[62183]: INFO nova.compute.manager [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Terminating instance [ 1011.153997] env[62183]: DEBUG nova.compute.manager [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1011.154168] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1011.155051] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc62511b-b883-4138-8408-4e57e7233132 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.165454] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1011.169298] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de9fb8b8-f8fd-42bf-bf29-81503cbba45b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.182086] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387427, 'name': Rename_Task, 'duration_secs': 0.14181} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.183696] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1011.184099] env[62183]: DEBUG oslo_vmware.api [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1011.184099] env[62183]: value = "task-1387430" [ 1011.184099] env[62183]: _type = "Task" [ 1011.184099] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.184361] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6971e340-39eb-4f65-a335-59efe7b0db4b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.205123] env[62183]: DEBUG oslo_vmware.api [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387430, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.207714] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1011.207714] env[62183]: value = "task-1387431" [ 1011.207714] env[62183]: _type = "Task" [ 1011.207714] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.215654] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387431, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.265690] env[62183]: DEBUG oslo_concurrency.lockutils [None req-150c7abd-261f-4a25-9ce6-ebddd8677bc8 tempest-AttachVolumeShelveTestJSON-154320191 tempest-AttachVolumeShelveTestJSON-154320191-project-member] Lock "fb348784-62a0-4d1f-ac7f-f176f3da0dd9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.382s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.276249] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a99773d1-2546-4e07-a92e-9506b82247fc tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "a948464b-63aa-4bc8-9885-228049e96d37" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.706s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.412732] env[62183]: DEBUG oslo_vmware.api [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387429, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132238} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.413418] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.413418] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.413552] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.413962] env[62183]: INFO nova.compute.manager [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1011.414176] env[62183]: DEBUG oslo.service.loopingcall [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.415261] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Volume attach. Driver type: vmdk {{(pid=62183) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1011.415519] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294514', 'volume_id': '786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'name': 'volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c4b68d6-fedd-408f-a449-aace7400014a', 'attached_at': '', 'detached_at': '', 'volume_id': '786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'serial': '786b9d04-6ed3-4cb1-a773-6db867ff2adb'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1011.416142] env[62183]: DEBUG nova.compute.manager [-] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1011.416142] env[62183]: DEBUG nova.network.neutron [-] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1011.418422] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1876e02-fa04-4d44-bf91-6bb562fcedfb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.437865] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0744473-3a2f-441c-98c4-24f94c228c84 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.463398] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb/volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.463761] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6ac1050-7117-45e6-a824-f450b91c708c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.482966] env[62183]: DEBUG oslo_vmware.api [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1011.482966] env[62183]: value = "task-1387432" [ 1011.482966] env[62183]: _type = "Task" [ 1011.482966] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.492507] env[62183]: DEBUG oslo_vmware.api [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387432, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.696140] env[62183]: DEBUG oslo_vmware.api [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387430, 'name': PowerOffVM_Task, 'duration_secs': 0.28218} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.696465] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1011.696643] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1011.696963] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-013d43af-08d6-4a55-9004-148913e8dfe6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.716903] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387431, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.734344] env[62183]: DEBUG oslo_concurrency.lockutils [None req-58a38c13-e58c-4e80-a389-a651192eddbc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.735573] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.957s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.735809] env[62183]: DEBUG nova.objects.instance [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'resources' on Instance uuid d8817d24-b0cb-4956-b195-cc417ae09fb4 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.773391] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1011.773618] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1011.773812] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleting the datastore file [datastore2] ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.774250] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3975718c-acb8-4ccc-833a-f3699bddd1a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.780749] env[62183]: DEBUG oslo_vmware.api [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1011.780749] env[62183]: value = "task-1387434" [ 1011.780749] env[62183]: _type = "Task" [ 1011.780749] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.789431] env[62183]: DEBUG oslo_vmware.api [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387434, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.993832] env[62183]: DEBUG oslo_vmware.api [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.030278] env[62183]: DEBUG nova.compute.manager [req-e9c472ec-2a15-4b0f-bf96-f6c2e28def4b req-c3fa85dc-aeb2-4bd7-a69d-1a24f352de04 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Received event network-vif-deleted-21c40087-5244-413a-ae18-e7e27f2abc34 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.030493] env[62183]: INFO nova.compute.manager [req-e9c472ec-2a15-4b0f-bf96-f6c2e28def4b req-c3fa85dc-aeb2-4bd7-a69d-1a24f352de04 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Neutron deleted interface 21c40087-5244-413a-ae18-e7e27f2abc34; detaching it from the instance and deleting it from the info cache [ 1012.030703] env[62183]: DEBUG nova.network.neutron [req-e9c472ec-2a15-4b0f-bf96-f6c2e28def4b req-c3fa85dc-aeb2-4bd7-a69d-1a24f352de04 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.220903] env[62183]: DEBUG oslo_vmware.api [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387431, 'name': PowerOnVM_Task, 'duration_secs': 0.914048} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.223047] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1012.223047] env[62183]: INFO nova.compute.manager [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Took 10.44 seconds to spawn the instance on the hypervisor. [ 1012.223047] env[62183]: DEBUG nova.compute.manager [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1012.223444] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4bf5fb-7a20-4044-8725-c2cebe834a5c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.300068] env[62183]: DEBUG oslo_vmware.api [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387434, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.454818} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.300402] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.300608] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1012.300851] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1012.301044] env[62183]: INFO nova.compute.manager [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1012.301381] env[62183]: DEBUG oslo.service.loopingcall [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.304146] env[62183]: DEBUG nova.compute.manager [-] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1012.304264] env[62183]: DEBUG nova.network.neutron [-] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1012.455831] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8474e8ac-89b2-4fa5-834e-2f4881efd9de {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.470768] env[62183]: DEBUG nova.network.neutron [-] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.477053] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd1f5138-3bdf-42fc-8993-0caf6485e225 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.534757] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec09176a-d98e-4fba-86b3-9c92cb3e7c86 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.539908] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-215c5aaa-1904-4d8a-bfb7-f274bd96480f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.551099] env[62183]: DEBUG oslo_vmware.api [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387432, 'name': ReconfigVM_Task, 'duration_secs': 0.846378} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.555370] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Reconfigured VM instance instance-0000005a to attach disk [datastore1] volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb/volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.563766] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd09ceb5-6309-4e6b-8e60-d7c2797a55ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.584020] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42b2ee7-2d0d-4080-b76a-ea5db28d297b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.597021] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3fd414-d7c1-46d5-85f6-f8cf6a9cc55b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.629673] env[62183]: DEBUG nova.compute.provider_tree [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.645380] env[62183]: DEBUG nova.compute.manager [req-e9c472ec-2a15-4b0f-bf96-f6c2e28def4b req-c3fa85dc-aeb2-4bd7-a69d-1a24f352de04 service nova] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Detach interface failed, port_id=21c40087-5244-413a-ae18-e7e27f2abc34, reason: Instance 3a3ca703-2883-4aa9-a33f-326cc25d4838 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1012.646708] env[62183]: DEBUG oslo_vmware.api [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1012.646708] env[62183]: value = "task-1387435" [ 1012.646708] env[62183]: _type = "Task" [ 1012.646708] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.663231] env[62183]: DEBUG oslo_vmware.api [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387435, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.734539] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "88846abe-9e6e-4a88-bf6e-9978610213da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.734824] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.748868] env[62183]: INFO nova.compute.manager [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Took 21.26 seconds to build instance. [ 1012.982117] env[62183]: INFO nova.compute.manager [-] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Took 1.57 seconds to deallocate network for instance. [ 1013.151283] env[62183]: DEBUG nova.scheduler.client.report [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1013.168150] env[62183]: DEBUG oslo_vmware.api [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387435, 'name': ReconfigVM_Task, 'duration_secs': 0.206195} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.168150] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294514', 'volume_id': '786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'name': 'volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c4b68d6-fedd-408f-a449-aace7400014a', 'attached_at': '', 'detached_at': '', 'volume_id': '786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'serial': '786b9d04-6ed3-4cb1-a773-6db867ff2adb'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1013.241859] env[62183]: DEBUG nova.compute.manager [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1013.251736] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b94bf858-b4c4-40a7-8554-cb95ddf47abb tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.771s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.493814] env[62183]: DEBUG oslo_concurrency.lockutils [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.660109] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.924s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.666508] env[62183]: DEBUG oslo_concurrency.lockutils [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.169s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.666508] env[62183]: DEBUG nova.objects.instance [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lazy-loading 'resources' on Instance uuid 3a3ca703-2883-4aa9-a33f-326cc25d4838 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.691758] env[62183]: INFO nova.scheduler.client.report [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted allocations for instance d8817d24-b0cb-4956-b195-cc417ae09fb4 [ 1013.718249] env[62183]: DEBUG nova.network.neutron [-] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.766984] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.079089] env[62183]: DEBUG nova.compute.manager [req-c65ce240-3a87-4c75-91d2-eb50d3fb5902 req-17c30962-fdae-4248-8aa9-90f3a17cc3fa service nova] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Received event network-vif-deleted-8d0bfbea-f6ad-41ee-a0f1-4ccad145901f {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.206208] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9a133d5a-3577-4b8b-bd0b-e9d0f3853fce tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "d8817d24-b0cb-4956-b195-cc417ae09fb4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.457s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.225176] env[62183]: DEBUG nova.objects.instance [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.227012] env[62183]: INFO nova.compute.manager [-] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Took 1.92 seconds to deallocate network for instance. [ 1014.331189] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405fb77c-5475-4527-b4b0-564054099bcb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.339227] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a6fe92-3521-48d5-beb3-2c5cb6535da7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.380899] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a623007-de6b-49b9-bc26-83fae73816c9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.391376] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37625737-6310-44db-b051-2d5b1fb88cb7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.409650] env[62183]: DEBUG nova.compute.provider_tree [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.733461] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ab447aac-bb25-401a-992c-f733dbe13303 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.932s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.740112] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.912638] env[62183]: DEBUG nova.scheduler.client.report [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1015.282067] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "a5018b4e-016d-4c61-bfbf-2c039ab79499" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.282067] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.282067] env[62183]: DEBUG nova.compute.manager [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.282067] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d314f35-97b0-4aa0-95d8-5495ad61f036 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.289369] env[62183]: DEBUG nova.compute.manager [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62183) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1015.290196] env[62183]: DEBUG nova.objects.instance [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lazy-loading 'flavor' on Instance uuid a5018b4e-016d-4c61-bfbf-2c039ab79499 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.330862] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "8c4b68d6-fedd-408f-a449-aace7400014a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.331162] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.331357] env[62183]: DEBUG nova.compute.manager [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.332293] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3c6a10-4426-4a3d-8c7c-d94b30fbdca7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.341719] env[62183]: DEBUG nova.compute.manager [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62183) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1015.341719] env[62183]: DEBUG nova.objects.instance [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.425207] env[62183]: DEBUG oslo_concurrency.lockutils [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.428156] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.661s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.431253] env[62183]: INFO nova.compute.claims [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1015.452383] env[62183]: INFO nova.scheduler.client.report [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted allocations for instance 3a3ca703-2883-4aa9-a33f-326cc25d4838 [ 1015.992346] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.993472] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-032e0228-1759-4ad9-9a9a-072ad7ac2382 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.002179] env[62183]: DEBUG oslo_vmware.api [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1016.002179] env[62183]: value = "task-1387436" [ 1016.002179] env[62183]: _type = "Task" [ 1016.002179] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.017136] env[62183]: DEBUG oslo_vmware.api [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387436, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.019435] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1016.022372] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2193f2a-e4dd-4712-bab7-f6c0caea0577 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.027806] env[62183]: DEBUG oslo_vmware.api [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1016.027806] env[62183]: value = "task-1387437" [ 1016.027806] env[62183]: _type = "Task" [ 1016.027806] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.042195] env[62183]: DEBUG oslo_vmware.api [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387437, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.460871] env[62183]: DEBUG oslo_concurrency.lockutils [None req-745a7422-3ffc-409e-b8fe-3a028a17868f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "3a3ca703-2883-4aa9-a33f-326cc25d4838" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.172s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.514417] env[62183]: DEBUG oslo_vmware.api [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387436, 'name': PowerOffVM_Task, 'duration_secs': 0.215966} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.514756] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1016.515226] env[62183]: DEBUG nova.compute.manager [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.515777] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b916a0d-8d1c-422f-b85e-8fc4e948bc1b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.540992] env[62183]: DEBUG oslo_vmware.api [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387437, 'name': PowerOffVM_Task, 'duration_secs': 0.281896} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.540992] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1016.541190] env[62183]: DEBUG nova.compute.manager [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.542773] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ffdc11-1cac-4f25-a241-f246f176aa44 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.637752] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7431767d-85a0-4261-815e-4c50e6387be9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.646969] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7a6be7-277a-4c0d-9679-06d4fcc77eee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.677244] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c0aaef-bf48-4839-a21b-818a14ad1f01 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.684829] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aae325a-70b9-45fa-b24d-7669c19c9c33 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.697819] env[62183]: DEBUG nova.compute.provider_tree [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.834009] env[62183]: DEBUG nova.compute.manager [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Stashing vm_state: active {{(pid=62183) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1017.035790] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6354f680-dea4-4a65-bdce-04b3676c8fc2 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.756s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.056289] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e03c3b7b-c12f-40ab-9a68-ce86c98ac68b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.725s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.201646] env[62183]: DEBUG nova.scheduler.client.report [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1017.353901] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.607667] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquiring lock "f51495aa-32b3-429c-8421-65f0a2587ea8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.608065] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "f51495aa-32b3-429c-8421-65f0a2587ea8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.708240] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.708816] env[62183]: DEBUG nova.compute.manager [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1017.711799] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.972s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.712037] env[62183]: DEBUG nova.objects.instance [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'resources' on Instance uuid ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.960928] env[62183]: DEBUG nova.objects.instance [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.009764] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.010014] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.067333] env[62183]: DEBUG nova.objects.instance [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lazy-loading 'flavor' on Instance uuid a5018b4e-016d-4c61-bfbf-2c039ab79499 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.110750] env[62183]: DEBUG nova.compute.manager [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1018.216053] env[62183]: DEBUG nova.compute.utils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1018.220260] env[62183]: DEBUG nova.compute.manager [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1018.220455] env[62183]: DEBUG nova.network.neutron [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1018.261444] env[62183]: DEBUG nova.policy [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4b92513f568466e81075af3fa4604fa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88022032e6e04a4f96bc49c2ca5ede29', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1018.377018] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e39b74-2ead-4143-969f-5bb0c1556de8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.383602] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d77f77-4553-47ac-8771-288785a21b89 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.417982] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a5a89a-d707-47bf-9ab2-4953bf949055 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.425748] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19578d91-1b05-46c2-8d6a-89df8832dca6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.442298] env[62183]: DEBUG nova.compute.provider_tree [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.465589] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.465759] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.466093] env[62183]: DEBUG nova.network.neutron [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1018.466141] env[62183]: DEBUG nova.objects.instance [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'info_cache' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.512226] env[62183]: DEBUG nova.compute.manager [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1018.572449] env[62183]: DEBUG oslo_concurrency.lockutils [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "refresh_cache-a5018b4e-016d-4c61-bfbf-2c039ab79499" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.572636] env[62183]: DEBUG oslo_concurrency.lockutils [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquired lock "refresh_cache-a5018b4e-016d-4c61-bfbf-2c039ab79499" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.572821] env[62183]: DEBUG nova.network.neutron [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1018.573043] env[62183]: DEBUG nova.objects.instance [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lazy-loading 'info_cache' on Instance uuid a5018b4e-016d-4c61-bfbf-2c039ab79499 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.583112] env[62183]: DEBUG nova.network.neutron [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Successfully created port: be785bab-5120-4093-b24c-25e5c14608bf {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1018.630475] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.721433] env[62183]: DEBUG nova.compute.manager [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1018.946007] env[62183]: DEBUG nova.scheduler.client.report [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1018.969697] env[62183]: DEBUG nova.objects.base [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Object Instance<8c4b68d6-fedd-408f-a449-aace7400014a> lazy-loaded attributes: flavor,info_cache {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1019.031011] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.076298] env[62183]: DEBUG nova.objects.base [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1019.451128] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.739s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.454275] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.100s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.469325] env[62183]: INFO nova.scheduler.client.report [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleted allocations for instance ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de [ 1019.731278] env[62183]: DEBUG nova.compute.manager [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1019.757340] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1019.757619] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1019.757783] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.757968] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1019.758507] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.758738] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1019.759015] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1019.759282] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1019.759448] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1019.759661] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1019.759885] env[62183]: DEBUG nova.virt.hardware [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1019.760832] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493972f8-c75d-4ff0-a425-335fd2132b8a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.764849] env[62183]: DEBUG nova.network.neutron [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updating instance_info_cache with network_info: [{"id": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "address": "fa:16:3e:10:9e:d2", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape301232c-2b", "ovs_interfaceid": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.773720] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a1df83-d888-4afa-aaa1-5d168e857fcb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.873800] env[62183]: DEBUG nova.network.neutron [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Updating instance_info_cache with network_info: [{"id": "f48d3c73-edab-4cf3-94cc-498503b5c465", "address": "fa:16:3e:65:a7:94", "network": {"id": "7cbe5832-931e-430d-95f5-8f48781f095b", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-236543802-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "587870b99b844c819d27ac4852e44c3d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e55c248-c504-4c7a-bbe9-f42cf417aee7", "external-id": "nsx-vlan-transportzone-471", "segmentation_id": 471, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf48d3c73-ed", "ovs_interfaceid": "f48d3c73-edab-4cf3-94cc-498503b5c465", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.958685] env[62183]: INFO nova.compute.claims [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.978144] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2d41dfe6-f06c-411e-a5d0-05a88624512c tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.830s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.170425] env[62183]: DEBUG nova.compute.manager [req-1990721a-2da2-4d31-9bce-d95ca69f82b8 req-76dafb06-aea4-4415-baf9-9d97b1c6a4d9 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Received event network-vif-plugged-be785bab-5120-4093-b24c-25e5c14608bf {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.170669] env[62183]: DEBUG oslo_concurrency.lockutils [req-1990721a-2da2-4d31-9bce-d95ca69f82b8 req-76dafb06-aea4-4415-baf9-9d97b1c6a4d9 service nova] Acquiring lock "88846abe-9e6e-4a88-bf6e-9978610213da-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.170893] env[62183]: DEBUG oslo_concurrency.lockutils [req-1990721a-2da2-4d31-9bce-d95ca69f82b8 req-76dafb06-aea4-4415-baf9-9d97b1c6a4d9 service nova] Lock "88846abe-9e6e-4a88-bf6e-9978610213da-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.171181] env[62183]: DEBUG oslo_concurrency.lockutils [req-1990721a-2da2-4d31-9bce-d95ca69f82b8 req-76dafb06-aea4-4415-baf9-9d97b1c6a4d9 service nova] Lock "88846abe-9e6e-4a88-bf6e-9978610213da-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.171445] env[62183]: DEBUG nova.compute.manager [req-1990721a-2da2-4d31-9bce-d95ca69f82b8 req-76dafb06-aea4-4415-baf9-9d97b1c6a4d9 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] No waiting events found dispatching network-vif-plugged-be785bab-5120-4093-b24c-25e5c14608bf {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.171699] env[62183]: WARNING nova.compute.manager [req-1990721a-2da2-4d31-9bce-d95ca69f82b8 req-76dafb06-aea4-4415-baf9-9d97b1c6a4d9 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Received unexpected event network-vif-plugged-be785bab-5120-4093-b24c-25e5c14608bf for instance with vm_state building and task_state spawning. [ 1020.268520] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Releasing lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.294252] env[62183]: DEBUG nova.network.neutron [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Successfully updated port: be785bab-5120-4093-b24c-25e5c14608bf {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.376575] env[62183]: DEBUG oslo_concurrency.lockutils [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Releasing lock "refresh_cache-a5018b4e-016d-4c61-bfbf-2c039ab79499" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.465130] env[62183]: INFO nova.compute.resource_tracker [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating resource usage from migration e4b22d4f-df7b-4efe-a48e-ddc16cb72b96 [ 1020.629487] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c490d181-6eeb-468c-a89f-f2fe2306e8dc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.637754] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f59eb8-d6a8-4209-8a95-0d01a97cdc35 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.668425] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7029d27-435d-4177-b266-dc4e99791587 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.675564] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3cfe2d-6012-47bc-80a7-d20e36b3dd2d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.688891] env[62183]: DEBUG nova.compute.provider_tree [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.773451] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1020.773771] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-948c45bf-fd6d-4fb0-b267-be541b9cbd58 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.781101] env[62183]: DEBUG oslo_vmware.api [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1020.781101] env[62183]: value = "task-1387438" [ 1020.781101] env[62183]: _type = "Task" [ 1020.781101] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.789282] env[62183]: DEBUG oslo_vmware.api [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387438, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.797945] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.798106] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.798243] env[62183]: DEBUG nova.network.neutron [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1020.880578] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1020.880942] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30863204-3680-4b8f-ac8a-0d557a586155 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.890094] env[62183]: DEBUG oslo_vmware.api [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1020.890094] env[62183]: value = "task-1387439" [ 1020.890094] env[62183]: _type = "Task" [ 1020.890094] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.903997] env[62183]: DEBUG oslo_vmware.api [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387439, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.192424] env[62183]: DEBUG nova.scheduler.client.report [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1021.295831] env[62183]: DEBUG oslo_vmware.api [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387438, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.345923] env[62183]: DEBUG nova.network.neutron [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1021.399621] env[62183]: DEBUG oslo_vmware.api [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387439, 'name': PowerOnVM_Task, 'duration_secs': 0.402309} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.399941] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.400222] env[62183]: DEBUG nova.compute.manager [None req-78320c7d-f568-46c9-82ef-692cc29fb019 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.401086] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830b7d4d-942a-4957-a09c-9c9afb6b4914 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.501135] env[62183]: DEBUG nova.network.neutron [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance_info_cache with network_info: [{"id": "be785bab-5120-4093-b24c-25e5c14608bf", "address": "fa:16:3e:da:b3:0d", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe785bab-51", "ovs_interfaceid": "be785bab-5120-4093-b24c-25e5c14608bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.700024] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.246s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.700288] env[62183]: INFO nova.compute.manager [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Migrating [ 1021.706594] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.076s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.708021] env[62183]: INFO nova.compute.claims [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.792708] env[62183]: DEBUG oslo_vmware.api [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387438, 'name': PowerOnVM_Task, 'duration_secs': 0.527309} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.792996] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.793364] env[62183]: DEBUG nova.compute.manager [None req-5a543c49-0216-4ea2-8931-e0cdf38cb7c0 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.794103] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cc7ec5-f20a-4387-8a46-a6c4efe170b2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.895407] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "1c180169-ebb5-4870-8e9d-925fe107b707" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.895657] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.004219] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.004778] env[62183]: DEBUG nova.compute.manager [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Instance network_info: |[{"id": "be785bab-5120-4093-b24c-25e5c14608bf", "address": "fa:16:3e:da:b3:0d", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe785bab-51", "ovs_interfaceid": "be785bab-5120-4093-b24c-25e5c14608bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1022.005198] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:b3:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ff90ec9-3c7e-4e76-b409-fcf37fc588d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be785bab-5120-4093-b24c-25e5c14608bf', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.012448] env[62183]: DEBUG oslo.service.loopingcall [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.012886] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1022.012886] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6b638c4-9e39-4f2a-a998-8dbb4b413bed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.033796] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.033796] env[62183]: value = "task-1387440" [ 1022.033796] env[62183]: _type = "Task" [ 1022.033796] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.041750] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387440, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.191865] env[62183]: DEBUG nova.compute.manager [req-e61dc88f-9c72-466d-8e90-ae102b078305 req-e484445b-7470-4562-88fb-01910c6c7024 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Received event network-changed-be785bab-5120-4093-b24c-25e5c14608bf {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.191865] env[62183]: DEBUG nova.compute.manager [req-e61dc88f-9c72-466d-8e90-ae102b078305 req-e484445b-7470-4562-88fb-01910c6c7024 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Refreshing instance network info cache due to event network-changed-be785bab-5120-4093-b24c-25e5c14608bf. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1022.192197] env[62183]: DEBUG oslo_concurrency.lockutils [req-e61dc88f-9c72-466d-8e90-ae102b078305 req-e484445b-7470-4562-88fb-01910c6c7024 service nova] Acquiring lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.192197] env[62183]: DEBUG oslo_concurrency.lockutils [req-e61dc88f-9c72-466d-8e90-ae102b078305 req-e484445b-7470-4562-88fb-01910c6c7024 service nova] Acquired lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.192287] env[62183]: DEBUG nova.network.neutron [req-e61dc88f-9c72-466d-8e90-ae102b078305 req-e484445b-7470-4562-88fb-01910c6c7024 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Refreshing network info cache for port be785bab-5120-4093-b24c-25e5c14608bf {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1022.219014] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.219230] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.219465] env[62183]: DEBUG nova.network.neutron [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1022.397823] env[62183]: DEBUG nova.compute.manager [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1022.544832] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387440, 'name': CreateVM_Task, 'duration_secs': 0.365904} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.545139] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1022.545841] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.546030] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.546379] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1022.546643] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb150a92-d4fb-487e-a802-d2c8a9d9c929 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.551225] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1022.551225] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5272f361-dee2-062f-8084-fa43a08f23c9" [ 1022.551225] env[62183]: _type = "Task" [ 1022.551225] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.559287] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5272f361-dee2-062f-8084-fa43a08f23c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.893377] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ec01d1-22af-410d-896c-8be0d86ea18e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.909384] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6a5ffb-9367-4527-85d6-91856485cd2f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.947336] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.948206] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d544cf1e-6783-429d-abb2-a3dc7d2b8f70 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.956198] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8425a0-0960-4f4a-b69e-0db4403b1527 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.969780] env[62183]: DEBUG nova.compute.provider_tree [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.062460] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5272f361-dee2-062f-8084-fa43a08f23c9, 'name': SearchDatastore_Task, 'duration_secs': 0.009745} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.062896] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.062896] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.063099] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.063265] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.063455] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.063743] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c707197-84a2-4593-a198-742de5b4ac2e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.075139] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.075139] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1023.075282] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72a17faf-75c0-42af-830a-652fe234bdae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.078528] env[62183]: DEBUG nova.network.neutron [req-e61dc88f-9c72-466d-8e90-ae102b078305 req-e484445b-7470-4562-88fb-01910c6c7024 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updated VIF entry in instance network info cache for port be785bab-5120-4093-b24c-25e5c14608bf. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1023.078911] env[62183]: DEBUG nova.network.neutron [req-e61dc88f-9c72-466d-8e90-ae102b078305 req-e484445b-7470-4562-88fb-01910c6c7024 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance_info_cache with network_info: [{"id": "be785bab-5120-4093-b24c-25e5c14608bf", "address": "fa:16:3e:da:b3:0d", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe785bab-51", "ovs_interfaceid": "be785bab-5120-4093-b24c-25e5c14608bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.081494] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1023.081494] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e58c0c-da78-89f1-fb88-0bbbc3988092" [ 1023.081494] env[62183]: _type = "Task" [ 1023.081494] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.089396] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e58c0c-da78-89f1-fb88-0bbbc3988092, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.112971] env[62183]: DEBUG nova.network.neutron [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.473118] env[62183]: DEBUG nova.scheduler.client.report [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1023.582989] env[62183]: DEBUG oslo_concurrency.lockutils [req-e61dc88f-9c72-466d-8e90-ae102b078305 req-e484445b-7470-4562-88fb-01910c6c7024 service nova] Releasing lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.594139] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e58c0c-da78-89f1-fb88-0bbbc3988092, 'name': SearchDatastore_Task, 'duration_secs': 0.009728} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.595195] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-857f5bf0-806a-46f9-b19e-10cd731f8d0e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.603531] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1023.603531] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52ff532d-ed13-1e44-4e86-6e04fb846ca9" [ 1023.603531] env[62183]: _type = "Task" [ 1023.603531] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.611611] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ff532d-ed13-1e44-4e86-6e04fb846ca9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.615325] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.978730] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.272s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.979317] env[62183]: DEBUG nova.compute.manager [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1023.982335] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.951s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.983723] env[62183]: INFO nova.compute.claims [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.122712] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52ff532d-ed13-1e44-4e86-6e04fb846ca9, 'name': SearchDatastore_Task, 'duration_secs': 0.010003} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.124054] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.124054] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 88846abe-9e6e-4a88-bf6e-9978610213da/88846abe-9e6e-4a88-bf6e-9978610213da.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1024.124993] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff3182c2-1dc9-450a-b309-d1478b6cc2fb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.133426] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1024.133426] env[62183]: value = "task-1387441" [ 1024.133426] env[62183]: _type = "Task" [ 1024.133426] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.141997] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387441, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.488440] env[62183]: DEBUG nova.compute.utils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.491432] env[62183]: DEBUG nova.compute.manager [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1024.493105] env[62183]: DEBUG nova.network.neutron [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1024.533657] env[62183]: DEBUG nova.policy [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '847c593736d84610bd9459c0b5598c03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc61446783c74acd8f66d4477b3a9331', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.589671] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.590152] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.590513] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.590843] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.591153] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.593678] env[62183]: INFO nova.compute.manager [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Terminating instance [ 1024.595662] env[62183]: DEBUG nova.compute.manager [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1024.595924] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.596796] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8cc769-1bbe-47a3-a65a-fb3c279a465d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.604713] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.604915] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f7db3e9-b378-4b91-a827-c6252eefc499 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.610698] env[62183]: DEBUG oslo_vmware.api [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1024.610698] env[62183]: value = "task-1387442" [ 1024.610698] env[62183]: _type = "Task" [ 1024.610698] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.618791] env[62183]: DEBUG oslo_vmware.api [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387442, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.643748] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387441, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.803293] env[62183]: DEBUG nova.network.neutron [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Successfully created port: 1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1024.995297] env[62183]: DEBUG nova.compute.manager [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1025.122732] env[62183]: DEBUG oslo_vmware.api [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387442, 'name': PowerOffVM_Task, 'duration_secs': 0.320347} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.125107] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.125303] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.125770] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24c47466-8475-4d24-8974-97f8a38460a0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.131606] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a055da7f-3de0-4f21-aa71-a0f79f19e92a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.152752] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance 'a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a' progress to 0 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1025.164807] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387441, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.874295} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.164807] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 88846abe-9e6e-4a88-bf6e-9978610213da/88846abe-9e6e-4a88-bf6e-9978610213da.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1025.164807] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1025.164807] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7cfac8b1-6779-460d-af59-68b7d45d2e8b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.170492] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1025.170492] env[62183]: value = "task-1387444" [ 1025.170492] env[62183]: _type = "Task" [ 1025.170492] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.174841] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea09828-6023-46e4-8fe4-9a105b2af52f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.183987] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387444, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.186660] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5e5385-f8ac-4445-a535-40aa927531d9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.218349] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c460c0-9649-4c00-b3fb-1039b17b495a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.220992] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.221204] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.221390] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Deleting the datastore file [datastore1] ac58afb9-9f4c-4092-bcfc-b3ed1433a356 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.221902] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e7d9269f-5d25-4e53-b511-d09d2c7f337a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.227874] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318ff014-4104-4deb-bb51-3d469fd72689 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.232723] env[62183]: DEBUG oslo_vmware.api [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1025.232723] env[62183]: value = "task-1387445" [ 1025.232723] env[62183]: _type = "Task" [ 1025.232723] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.245225] env[62183]: DEBUG nova.compute.provider_tree [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.251332] env[62183]: DEBUG oslo_vmware.api [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387445, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.663678] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.664106] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d2ef8a1-db61-469a-bc03-45289c6a0447 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.672076] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1025.672076] env[62183]: value = "task-1387446" [ 1025.672076] env[62183]: _type = "Task" [ 1025.672076] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.686739] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387444, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.184222} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.691016] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1025.691439] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387446, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.692525] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9f62f0-9164-4611-8537-5dc4c5a8ce2c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.723891] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 88846abe-9e6e-4a88-bf6e-9978610213da/88846abe-9e6e-4a88-bf6e-9978610213da.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.724297] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7d51aed-c23f-498f-94ab-386f7efd3a1e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.751757] env[62183]: DEBUG nova.scheduler.client.report [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1025.757988] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1025.757988] env[62183]: value = "task-1387447" [ 1025.757988] env[62183]: _type = "Task" [ 1025.757988] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.766715] env[62183]: DEBUG oslo_vmware.api [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387445, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.452035} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.767518] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.767768] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1025.768024] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1025.768271] env[62183]: INFO nova.compute.manager [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1025.768593] env[62183]: DEBUG oslo.service.loopingcall [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.769569] env[62183]: DEBUG nova.compute.manager [-] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1025.769701] env[62183]: DEBUG nova.network.neutron [-] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1025.776195] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387447, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.006674] env[62183]: DEBUG nova.compute.manager [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1026.009501] env[62183]: DEBUG nova.compute.manager [req-c2fe8dd3-f902-4baf-8b78-03a6c86a7fa9 req-4de0c0a0-e061-4dcf-bdfd-a028d9f3af19 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Received event network-vif-deleted-5034c986-efd5-4de5-8bc7-6391f23605cd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.009705] env[62183]: INFO nova.compute.manager [req-c2fe8dd3-f902-4baf-8b78-03a6c86a7fa9 req-4de0c0a0-e061-4dcf-bdfd-a028d9f3af19 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Neutron deleted interface 5034c986-efd5-4de5-8bc7-6391f23605cd; detaching it from the instance and deleting it from the info cache [ 1026.009877] env[62183]: DEBUG nova.network.neutron [req-c2fe8dd3-f902-4baf-8b78-03a6c86a7fa9 req-4de0c0a0-e061-4dcf-bdfd-a028d9f3af19 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.033883] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.034155] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.034322] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.034506] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.034656] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.034806] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.035048] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.035251] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.035431] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.035786] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.035786] env[62183]: DEBUG nova.virt.hardware [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.036648] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0984f07-9c5f-4084-b969-ca2c88bd7011 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.044376] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf54618-9941-4d85-a418-ea4ba94da4c0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.188697] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387446, 'name': PowerOffVM_Task, 'duration_secs': 0.199421} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.188973] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1026.189186] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance 'a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a' progress to 17 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1026.260027] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.260604] env[62183]: DEBUG nova.compute.manager [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1026.263185] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.316s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.264624] env[62183]: INFO nova.compute.claims [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.278033] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387447, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.487684] env[62183]: DEBUG nova.network.neutron [-] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.512370] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f62c96a-8c06-4061-894d-c6c5eccf32de {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.521676] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1345753-8de6-45d4-b9af-818badc3e791 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.547502] env[62183]: DEBUG nova.compute.manager [req-c2fe8dd3-f902-4baf-8b78-03a6c86a7fa9 req-4de0c0a0-e061-4dcf-bdfd-a028d9f3af19 service nova] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Detach interface failed, port_id=5034c986-efd5-4de5-8bc7-6391f23605cd, reason: Instance ac58afb9-9f4c-4092-bcfc-b3ed1433a356 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1026.585960] env[62183]: DEBUG nova.network.neutron [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Successfully updated port: 1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1026.696719] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1026.697031] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1026.697237] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.697442] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1026.697597] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.697756] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1026.697967] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1026.698157] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1026.698365] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1026.698555] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1026.698737] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.703779] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e702c117-1cac-4a26-8399-5d94366c0771 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.719177] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1026.719177] env[62183]: value = "task-1387448" [ 1026.719177] env[62183]: _type = "Task" [ 1026.719177] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.728798] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387448, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.775015] env[62183]: DEBUG nova.compute.utils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.778478] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387447, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.778975] env[62183]: DEBUG nova.compute.manager [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1026.779163] env[62183]: DEBUG nova.network.neutron [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1026.823471] env[62183]: DEBUG nova.policy [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88efcb2057e84413806f9ed809c260e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eda7e297c551407295b3c67fa0f1dbc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1026.991012] env[62183]: INFO nova.compute.manager [-] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Took 1.22 seconds to deallocate network for instance. [ 1027.070460] env[62183]: DEBUG nova.network.neutron [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Successfully created port: f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1027.089281] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquiring lock "refresh_cache-f51495aa-32b3-429c-8421-65f0a2587ea8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.089432] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquired lock "refresh_cache-f51495aa-32b3-429c-8421-65f0a2587ea8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.089637] env[62183]: DEBUG nova.network.neutron [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.229598] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387448, 'name': ReconfigVM_Task, 'duration_secs': 0.192104} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.230369] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance 'a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a' progress to 33 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1027.278909] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387447, 'name': ReconfigVM_Task, 'duration_secs': 1.280004} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.279390] env[62183]: DEBUG nova.compute.manager [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1027.281868] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 88846abe-9e6e-4a88-bf6e-9978610213da/88846abe-9e6e-4a88-bf6e-9978610213da.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.284898] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37089957-081d-4d6b-a536-58805826d3b2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.292404] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1027.292404] env[62183]: value = "task-1387449" [ 1027.292404] env[62183]: _type = "Task" [ 1027.292404] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.301665] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387449, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.435840] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286fb147-8599-4b8f-9f19-15c1b276b225 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.442962] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8c6351-8d7e-45dc-ad97-32b1655967c9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.472852] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8e58e6-c240-4ee5-a36b-42a9d3e9d529 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.479894] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472a78bb-b217-4306-ba88-34d5202071fc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.492330] env[62183]: DEBUG nova.compute.provider_tree [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.497894] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.631719] env[62183]: DEBUG nova.network.neutron [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.737188] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1027.737590] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1027.737764] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1027.737959] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1027.738391] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1027.738617] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1027.739127] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1027.739341] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1027.739530] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1027.739704] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1027.739884] env[62183]: DEBUG nova.virt.hardware [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.745225] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Reconfiguring VM instance instance-00000045 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1027.747933] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87c379be-5802-4082-a39c-15fa025e6400 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.767930] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1027.767930] env[62183]: value = "task-1387450" [ 1027.767930] env[62183]: _type = "Task" [ 1027.767930] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.776160] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387450, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.799028] env[62183]: DEBUG nova.network.neutron [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Updating instance_info_cache with network_info: [{"id": "1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8", "address": "fa:16:3e:8c:28:ca", "network": {"id": "b4b8c84a-5d7f-43c3-bd09-3498994b2d9f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1800707143-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc61446783c74acd8f66d4477b3a9331", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77ccbd87-ecfd-4b2d-a1ea-29774addcef6", "external-id": "nsx-vlan-transportzone-385", "segmentation_id": 385, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1afd9c1a-ea", "ovs_interfaceid": "1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.808050] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387449, 'name': Rename_Task, 'duration_secs': 0.145663} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.808571] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1027.808941] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c67f2fd5-3979-40ec-a35d-a7796be5bcf5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.816168] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1027.816168] env[62183]: value = "task-1387451" [ 1027.816168] env[62183]: _type = "Task" [ 1027.816168] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.825120] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387451, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.995416] env[62183]: DEBUG nova.scheduler.client.report [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1028.032234] env[62183]: DEBUG nova.compute.manager [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Received event network-vif-plugged-1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.032477] env[62183]: DEBUG oslo_concurrency.lockutils [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] Acquiring lock "f51495aa-32b3-429c-8421-65f0a2587ea8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.032668] env[62183]: DEBUG oslo_concurrency.lockutils [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] Lock "f51495aa-32b3-429c-8421-65f0a2587ea8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.032839] env[62183]: DEBUG oslo_concurrency.lockutils [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] Lock "f51495aa-32b3-429c-8421-65f0a2587ea8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.033023] env[62183]: DEBUG nova.compute.manager [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] No waiting events found dispatching network-vif-plugged-1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1028.033201] env[62183]: WARNING nova.compute.manager [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Received unexpected event network-vif-plugged-1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8 for instance with vm_state building and task_state spawning. [ 1028.033359] env[62183]: DEBUG nova.compute.manager [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Received event network-changed-1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.033535] env[62183]: DEBUG nova.compute.manager [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Refreshing instance network info cache due to event network-changed-1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.033729] env[62183]: DEBUG oslo_concurrency.lockutils [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] Acquiring lock "refresh_cache-f51495aa-32b3-429c-8421-65f0a2587ea8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.278213] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387450, 'name': ReconfigVM_Task, 'duration_secs': 0.17464} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.278542] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Reconfigured VM instance instance-00000045 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1028.279229] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c0b3bc-2986-48e9-ae7d-1b39c0790599 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.295713] env[62183]: DEBUG nova.compute.manager [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1028.305229] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a/a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.305882] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Releasing lock "refresh_cache-f51495aa-32b3-429c-8421-65f0a2587ea8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.306204] env[62183]: DEBUG nova.compute.manager [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Instance network_info: |[{"id": "1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8", "address": "fa:16:3e:8c:28:ca", "network": {"id": "b4b8c84a-5d7f-43c3-bd09-3498994b2d9f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1800707143-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc61446783c74acd8f66d4477b3a9331", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77ccbd87-ecfd-4b2d-a1ea-29774addcef6", "external-id": "nsx-vlan-transportzone-385", "segmentation_id": 385, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1afd9c1a-ea", "ovs_interfaceid": "1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1028.306641] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3af558a3-51e1-445d-8bb2-ed831e2fb02c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.319151] env[62183]: DEBUG oslo_concurrency.lockutils [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] Acquired lock "refresh_cache-f51495aa-32b3-429c-8421-65f0a2587ea8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.319363] env[62183]: DEBUG nova.network.neutron [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Refreshing network info cache for port 1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.320761] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:28:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '77ccbd87-ecfd-4b2d-a1ea-29774addcef6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.328753] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Creating folder: Project (dc61446783c74acd8f66d4477b3a9331). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1028.331700] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0323cfff-7f29-4d4a-958e-f923fddab6d5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.342821] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387451, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.345966] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.346210] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.346373] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.346560] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.346711] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.346861] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.347155] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.347285] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.347462] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.347629] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.347803] env[62183]: DEBUG nova.virt.hardware [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.348157] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1028.348157] env[62183]: value = "task-1387452" [ 1028.348157] env[62183]: _type = "Task" [ 1028.348157] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.349115] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb11da51-df18-49fb-95a2-57f33378adef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.352577] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Created folder: Project (dc61446783c74acd8f66d4477b3a9331) in parent group-v294392. [ 1028.352760] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Creating folder: Instances. Parent ref: group-v294516. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1028.356228] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-afe4cdec-d013-4a04-9acb-d5da2d4462ac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.364346] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1acad20e-f902-467d-9106-2b7ceb01ea58 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.368665] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.369963] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Created folder: Instances in parent group-v294516. [ 1028.370208] env[62183]: DEBUG oslo.service.loopingcall [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.370732] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.370982] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75545592-0a78-42aa-ae11-b0c4d61b54d3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.400183] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.400183] env[62183]: value = "task-1387455" [ 1028.400183] env[62183]: _type = "Task" [ 1028.400183] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.408911] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387455, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.503134] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.238s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.503134] env[62183]: DEBUG nova.compute.manager [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1028.504931] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.007s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.509404] env[62183]: DEBUG nova.objects.instance [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lazy-loading 'resources' on Instance uuid ac58afb9-9f4c-4092-bcfc-b3ed1433a356 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.656272] env[62183]: DEBUG nova.network.neutron [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Successfully updated port: f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.840909] env[62183]: DEBUG oslo_vmware.api [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387451, 'name': PowerOnVM_Task, 'duration_secs': 0.557764} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.841213] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1028.841418] env[62183]: INFO nova.compute.manager [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Took 9.11 seconds to spawn the instance on the hypervisor. [ 1028.841596] env[62183]: DEBUG nova.compute.manager [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.842376] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5bb8ab-4e32-4fda-b1c1-ce086dd63b6e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.860345] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387452, 'name': ReconfigVM_Task, 'duration_secs': 0.369115} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.862459] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Reconfigured VM instance instance-00000045 to attach disk [datastore2] a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a/a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.862745] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance 'a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a' progress to 50 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1028.911806] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387455, 'name': CreateVM_Task, 'duration_secs': 0.32852} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.911977] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.912668] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.912847] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.913205] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1028.913467] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0954a74f-9442-4bd4-ac74-9e1a3d965b5d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.917957] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1028.917957] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f93eb5-4b8c-3c85-c6b4-c0af96bb357b" [ 1028.917957] env[62183]: _type = "Task" [ 1028.917957] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.925639] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f93eb5-4b8c-3c85-c6b4-c0af96bb357b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.012417] env[62183]: DEBUG nova.compute.utils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.016122] env[62183]: DEBUG nova.compute.manager [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1029.016323] env[62183]: DEBUG nova.network.neutron [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1029.076533] env[62183]: DEBUG nova.policy [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ae1837d5db145278417f7cdd55a3fea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06e81bd1a81d4009ae2a75fe819f9b7c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.079488] env[62183]: DEBUG nova.network.neutron [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Updated VIF entry in instance network info cache for port 1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.080137] env[62183]: DEBUG nova.network.neutron [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Updating instance_info_cache with network_info: [{"id": "1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8", "address": "fa:16:3e:8c:28:ca", "network": {"id": "b4b8c84a-5d7f-43c3-bd09-3498994b2d9f", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1800707143-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc61446783c74acd8f66d4477b3a9331", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "77ccbd87-ecfd-4b2d-a1ea-29774addcef6", "external-id": "nsx-vlan-transportzone-385", "segmentation_id": 385, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1afd9c1a-ea", "ovs_interfaceid": "1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.157953] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "refresh_cache-488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.157953] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "refresh_cache-488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.158068] env[62183]: DEBUG nova.network.neutron [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1029.160438] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329d7f80-2ac8-4fb1-a1fa-4f001ed66e42 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.168671] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984a1e5c-2353-40f7-81d2-d9a0cbb29bf9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.199885] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469b0dec-d766-4eba-b96c-87ac1f832ead {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.207316] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98ac1e9-58f2-4141-a6bb-7242d1af404e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.222557] env[62183]: DEBUG nova.compute.provider_tree [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.337243] env[62183]: DEBUG nova.network.neutron [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Successfully created port: 78754a06-2366-4beb-bc13-95b404d4acf6 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1029.360017] env[62183]: INFO nova.compute.manager [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Took 15.61 seconds to build instance. [ 1029.369038] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cd1d2f-cdec-4303-b5e2-c0d96b96e64b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.388416] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a9f3089-f4d9-40b5-8d22-fa5829a8663c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.406101] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance 'a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a' progress to 67 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1029.426852] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f93eb5-4b8c-3c85-c6b4-c0af96bb357b, 'name': SearchDatastore_Task, 'duration_secs': 0.009568} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.427767] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.427767] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.427767] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.427767] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.428036] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.428184] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-72a60c6e-4f57-458e-9ad6-cfee94ee371b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.436346] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.437688] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.437688] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c86975f6-6c73-44b5-8176-3749ff90554e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.444030] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1029.444030] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52e15745-4b2b-3e68-4a18-c87233cd1b0c" [ 1029.444030] env[62183]: _type = "Task" [ 1029.444030] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.451228] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e15745-4b2b-3e68-4a18-c87233cd1b0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.517508] env[62183]: DEBUG nova.compute.manager [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1029.583642] env[62183]: DEBUG oslo_concurrency.lockutils [req-b501dffb-16dc-43ca-8974-0e20d974324c req-8209ac70-9a57-4731-93ef-0ef2c03a42c3 service nova] Releasing lock "refresh_cache-f51495aa-32b3-429c-8421-65f0a2587ea8" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.692816] env[62183]: DEBUG nova.network.neutron [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1029.726081] env[62183]: DEBUG nova.scheduler.client.report [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1029.825152] env[62183]: DEBUG nova.network.neutron [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Updating instance_info_cache with network_info: [{"id": "f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75", "address": "fa:16:3e:b0:72:1d", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0b525f4-3e", "ovs_interfaceid": "f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.862354] env[62183]: DEBUG oslo_concurrency.lockutils [None req-af1662d3-98b0-4181-815d-044fc515ec3b tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.127s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.953430] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52e15745-4b2b-3e68-4a18-c87233cd1b0c, 'name': SearchDatastore_Task, 'duration_secs': 0.009247} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.953430] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ffa0994-c9a7-43b0-9f12-382cc69efe91 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.958281] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1029.958281] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]522d77c8-df64-9016-8ac9-102d311a702d" [ 1029.958281] env[62183]: _type = "Task" [ 1029.958281] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.966080] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]522d77c8-df64-9016-8ac9-102d311a702d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.968591] env[62183]: DEBUG nova.network.neutron [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Port bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e binding to destination host cpu-1 is already ACTIVE {{(pid=62183) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1030.059070] env[62183]: DEBUG nova.compute.manager [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Received event network-vif-plugged-f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.059548] env[62183]: DEBUG oslo_concurrency.lockutils [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] Acquiring lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.059879] env[62183]: DEBUG oslo_concurrency.lockutils [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] Lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.060297] env[62183]: DEBUG oslo_concurrency.lockutils [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] Lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.060593] env[62183]: DEBUG nova.compute.manager [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] No waiting events found dispatching network-vif-plugged-f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1030.060869] env[62183]: WARNING nova.compute.manager [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Received unexpected event network-vif-plugged-f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75 for instance with vm_state building and task_state spawning. [ 1030.061233] env[62183]: DEBUG nova.compute.manager [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Received event network-changed-f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.061506] env[62183]: DEBUG nova.compute.manager [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Refreshing instance network info cache due to event network-changed-f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.061795] env[62183]: DEBUG oslo_concurrency.lockutils [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] Acquiring lock "refresh_cache-488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.233393] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.728s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.261045] env[62183]: INFO nova.scheduler.client.report [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Deleted allocations for instance ac58afb9-9f4c-4092-bcfc-b3ed1433a356 [ 1030.327593] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "refresh_cache-488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.328144] env[62183]: DEBUG nova.compute.manager [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Instance network_info: |[{"id": "f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75", "address": "fa:16:3e:b0:72:1d", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0b525f4-3e", "ovs_interfaceid": "f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1030.328323] env[62183]: DEBUG oslo_concurrency.lockutils [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] Acquired lock "refresh_cache-488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.328586] env[62183]: DEBUG nova.network.neutron [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Refreshing network info cache for port f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1030.329826] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:72:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.338943] env[62183]: DEBUG oslo.service.loopingcall [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.339972] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.340305] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd9f2888-a1ec-406f-af22-4bbb34461df6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.360918] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.360918] env[62183]: value = "task-1387456" [ 1030.360918] env[62183]: _type = "Task" [ 1030.360918] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.368818] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387456, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.469591] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]522d77c8-df64-9016-8ac9-102d311a702d, 'name': SearchDatastore_Task, 'duration_secs': 0.00907} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.473623] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.473893] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] f51495aa-32b3-429c-8421-65f0a2587ea8/f51495aa-32b3-429c-8421-65f0a2587ea8.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.474206] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5ad828e-8d9a-41f8-8214-8a66419fc37e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.481571] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1030.481571] env[62183]: value = "task-1387457" [ 1030.481571] env[62183]: _type = "Task" [ 1030.481571] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.490232] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387457, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.527049] env[62183]: DEBUG nova.compute.manager [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1030.557321] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.557612] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.557814] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.558051] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.558140] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.558339] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.558925] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.558925] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.559053] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.559201] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.559325] env[62183]: DEBUG nova.virt.hardware [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.560280] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01ceaa0-37a3-4fbb-90b6-e87c13f62aac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.568424] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90e09ba-6be8-415f-befd-479e484bfb14 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.771577] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4e8f1b4c-4c8b-4708-9487-19e0d4782bc3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "ac58afb9-9f4c-4092-bcfc-b3ed1433a356" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.181s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.872117] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387456, 'name': CreateVM_Task, 'duration_secs': 0.295935} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.872326] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1030.873238] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.873416] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.873805] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1030.874139] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31f257dc-f932-40a5-a185-366770da70c3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.879527] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1030.879527] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52711e47-2468-9b82-b8ed-9bc34c070829" [ 1030.879527] env[62183]: _type = "Task" [ 1030.879527] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.888109] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52711e47-2468-9b82-b8ed-9bc34c070829, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.925156] env[62183]: DEBUG nova.network.neutron [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Successfully updated port: 78754a06-2366-4beb-bc13-95b404d4acf6 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1030.996570] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.996798] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.996974] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.003179] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387457, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449322} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.003706] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] f51495aa-32b3-429c-8421-65f0a2587ea8/f51495aa-32b3-429c-8421-65f0a2587ea8.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1031.003930] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1031.004214] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b5756bf-a8a5-4a81-839f-0fee2ba6c783 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.011315] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1031.011315] env[62183]: value = "task-1387458" [ 1031.011315] env[62183]: _type = "Task" [ 1031.011315] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.021326] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387458, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.131324] env[62183]: DEBUG nova.network.neutron [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Updated VIF entry in instance network info cache for port f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.131699] env[62183]: DEBUG nova.network.neutron [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Updating instance_info_cache with network_info: [{"id": "f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75", "address": "fa:16:3e:b0:72:1d", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0b525f4-3e", "ovs_interfaceid": "f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.209191] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "dbd34888-a333-40b4-ae9c-df5541b16704" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.209485] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "dbd34888-a333-40b4-ae9c-df5541b16704" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.209760] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "dbd34888-a333-40b4-ae9c-df5541b16704-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.209963] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "dbd34888-a333-40b4-ae9c-df5541b16704-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.210155] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "dbd34888-a333-40b4-ae9c-df5541b16704-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.212505] env[62183]: INFO nova.compute.manager [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Terminating instance [ 1031.214338] env[62183]: DEBUG nova.compute.manager [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1031.214546] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1031.215462] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42e168a-0c30-4f3e-922b-87b3c21f3cab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.223279] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.223519] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05f30a2f-a6f0-49ae-ad00-8ec68fd441db {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.229668] env[62183]: DEBUG oslo_vmware.api [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1031.229668] env[62183]: value = "task-1387459" [ 1031.229668] env[62183]: _type = "Task" [ 1031.229668] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.237494] env[62183]: DEBUG oslo_vmware.api [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387459, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.389637] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52711e47-2468-9b82-b8ed-9bc34c070829, 'name': SearchDatastore_Task, 'duration_secs': 0.039347} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.389949] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.390210] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.390458] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.390653] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.390797] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.391078] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c093c3f-11af-4a9c-9423-e98ffca407b7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.402231] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.402409] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.403115] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bac5afe0-ecff-44c2-a5d3-dc99664f95da {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.408333] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1031.408333] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5279fe73-ddf5-5715-014f-ec50dd589190" [ 1031.408333] env[62183]: _type = "Task" [ 1031.408333] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.415811] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5279fe73-ddf5-5715-014f-ec50dd589190, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.427591] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "refresh_cache-1c180169-ebb5-4870-8e9d-925fe107b707" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.427723] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "refresh_cache-1c180169-ebb5-4870-8e9d-925fe107b707" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.427873] env[62183]: DEBUG nova.network.neutron [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1031.521477] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387458, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064663} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.521819] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.522619] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c17757d-e5e9-4eb3-9f87-d73da490b6f3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.544774] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] f51495aa-32b3-429c-8421-65f0a2587ea8/f51495aa-32b3-429c-8421-65f0a2587ea8.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.545380] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0084d678-eddc-4a8a-9ee3-d1c74ab9954f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.566893] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1031.566893] env[62183]: value = "task-1387460" [ 1031.566893] env[62183]: _type = "Task" [ 1031.566893] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.575534] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387460, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.635020] env[62183]: DEBUG oslo_concurrency.lockutils [req-917734f0-6c02-4dbf-9cc0-24ebf9f73221 req-297405f9-68f1-4e97-8895-4acc8be5f940 service nova] Releasing lock "refresh_cache-488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.739167] env[62183]: DEBUG oslo_vmware.api [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387459, 'name': PowerOffVM_Task, 'duration_secs': 0.456236} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.739481] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.739654] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1031.739905] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b14cf8d4-67d8-4a00-9d76-02a23eb38460 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.806996] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1031.807397] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1031.807511] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Deleting the datastore file [datastore1] dbd34888-a333-40b4-ae9c-df5541b16704 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1031.807762] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f243784-0a29-4a2c-a4f0-68f4b2f3fea9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.814343] env[62183]: DEBUG oslo_vmware.api [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1031.814343] env[62183]: value = "task-1387462" [ 1031.814343] env[62183]: _type = "Task" [ 1031.814343] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.821650] env[62183]: DEBUG oslo_vmware.api [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387462, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.918222] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5279fe73-ddf5-5715-014f-ec50dd589190, 'name': SearchDatastore_Task, 'duration_secs': 0.00764} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.919062] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0059bc60-dcc5-4b42-bd30-34090c1949d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.924460] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1031.924460] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5247591f-c2f1-6795-3323-7b7231fd95b3" [ 1031.924460] env[62183]: _type = "Task" [ 1031.924460] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.933460] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5247591f-c2f1-6795-3323-7b7231fd95b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.972604] env[62183]: DEBUG nova.network.neutron [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.059894] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.060115] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.060311] env[62183]: DEBUG nova.network.neutron [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1032.076332] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387460, 'name': ReconfigVM_Task, 'duration_secs': 0.367108} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.076638] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Reconfigured VM instance instance-00000066 to attach disk [datastore1] f51495aa-32b3-429c-8421-65f0a2587ea8/f51495aa-32b3-429c-8421-65f0a2587ea8.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.078035] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c745cc0f-4cbb-4ad7-8204-e44178112ba1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.084182] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1032.084182] env[62183]: value = "task-1387463" [ 1032.084182] env[62183]: _type = "Task" [ 1032.084182] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.092548] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387463, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.113063] env[62183]: DEBUG nova.compute.manager [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Received event network-changed-be785bab-5120-4093-b24c-25e5c14608bf {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.113368] env[62183]: DEBUG nova.compute.manager [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Refreshing instance network info cache due to event network-changed-be785bab-5120-4093-b24c-25e5c14608bf. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.113481] env[62183]: DEBUG oslo_concurrency.lockutils [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] Acquiring lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.113684] env[62183]: DEBUG oslo_concurrency.lockutils [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] Acquired lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.113858] env[62183]: DEBUG nova.network.neutron [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Refreshing network info cache for port be785bab-5120-4093-b24c-25e5c14608bf {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1032.183845] env[62183]: DEBUG nova.network.neutron [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Updating instance_info_cache with network_info: [{"id": "78754a06-2366-4beb-bc13-95b404d4acf6", "address": "fa:16:3e:5f:09:88", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78754a06-23", "ovs_interfaceid": "78754a06-2366-4beb-bc13-95b404d4acf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.323718] env[62183]: DEBUG oslo_vmware.api [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387462, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158518} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.323968] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.324166] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1032.324351] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1032.324530] env[62183]: INFO nova.compute.manager [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1032.325115] env[62183]: DEBUG oslo.service.loopingcall [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.325115] env[62183]: DEBUG nova.compute.manager [-] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1032.325115] env[62183]: DEBUG nova.network.neutron [-] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1032.434666] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5247591f-c2f1-6795-3323-7b7231fd95b3, 'name': SearchDatastore_Task, 'duration_secs': 0.008714} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.434934] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.435162] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86/488d1c9b-fd4f-4b0d-bd05-9f42c2132b86.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1032.435452] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f52fac69-e30d-4407-93ae-158fb40b7a72 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.441051] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1032.441051] env[62183]: value = "task-1387464" [ 1032.441051] env[62183]: _type = "Task" [ 1032.441051] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.448351] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387464, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.595235] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387463, 'name': Rename_Task, 'duration_secs': 0.126737} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.598011] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1032.598331] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4fc04a8c-f82f-4b76-af3a-35432a40b445 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.605611] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1032.605611] env[62183]: value = "task-1387465" [ 1032.605611] env[62183]: _type = "Task" [ 1032.605611] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.614077] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387465, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.687315] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "refresh_cache-1c180169-ebb5-4870-8e9d-925fe107b707" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.687621] env[62183]: DEBUG nova.compute.manager [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Instance network_info: |[{"id": "78754a06-2366-4beb-bc13-95b404d4acf6", "address": "fa:16:3e:5f:09:88", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78754a06-23", "ovs_interfaceid": "78754a06-2366-4beb-bc13-95b404d4acf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1032.688088] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:09:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '205fb402-8eaf-4b61-8f57-8f216024179a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78754a06-2366-4beb-bc13-95b404d4acf6', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1032.695701] env[62183]: DEBUG oslo.service.loopingcall [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.695924] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1032.696169] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e91580ec-9df4-4308-8d33-2e5a7a7a122b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.719516] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1032.719516] env[62183]: value = "task-1387466" [ 1032.719516] env[62183]: _type = "Task" [ 1032.719516] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.728361] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387466, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.950694] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387464, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49039} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.950911] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86/488d1c9b-fd4f-4b0d-bd05-9f42c2132b86.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1032.951154] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.951420] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c602ae1-313f-4959-99e2-2d4e76fcc2dd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.956885] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1032.956885] env[62183]: value = "task-1387467" [ 1032.956885] env[62183]: _type = "Task" [ 1032.956885] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.964341] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387467, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.025747] env[62183]: DEBUG nova.network.neutron [-] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.030911] env[62183]: DEBUG nova.network.neutron [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.044925] env[62183]: DEBUG nova.network.neutron [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updated VIF entry in instance network info cache for port be785bab-5120-4093-b24c-25e5c14608bf. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1033.045336] env[62183]: DEBUG nova.network.neutron [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance_info_cache with network_info: [{"id": "be785bab-5120-4093-b24c-25e5c14608bf", "address": "fa:16:3e:da:b3:0d", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe785bab-51", "ovs_interfaceid": "be785bab-5120-4093-b24c-25e5c14608bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.115955] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387465, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.229977] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387466, 'name': CreateVM_Task, 'duration_secs': 0.341099} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.230186] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1033.230837] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.231018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.231414] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1033.231795] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8803116-fb43-4be7-9366-fbabc876d801 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.236584] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1033.236584] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5200cab2-3344-f3d4-2b1d-9f21c592fd8d" [ 1033.236584] env[62183]: _type = "Task" [ 1033.236584] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.246294] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5200cab2-3344-f3d4-2b1d-9f21c592fd8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.466354] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387467, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097496} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.466645] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1033.467455] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4ccf74-7c19-44de-a5b6-0e71dd828bb1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.489360] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86/488d1c9b-fd4f-4b0d-bd05-9f42c2132b86.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.489672] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6d5e759-9b7c-4b4d-8a15-71995726380c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.509895] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1033.509895] env[62183]: value = "task-1387468" [ 1033.509895] env[62183]: _type = "Task" [ 1033.509895] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.517658] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387468, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.528209] env[62183]: INFO nova.compute.manager [-] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Took 1.20 seconds to deallocate network for instance. [ 1033.533047] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.547886] env[62183]: DEBUG oslo_concurrency.lockutils [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] Releasing lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.548155] env[62183]: DEBUG nova.compute.manager [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Received event network-vif-plugged-78754a06-2366-4beb-bc13-95b404d4acf6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.548420] env[62183]: DEBUG oslo_concurrency.lockutils [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] Acquiring lock "1c180169-ebb5-4870-8e9d-925fe107b707-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.548653] env[62183]: DEBUG oslo_concurrency.lockutils [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] Lock "1c180169-ebb5-4870-8e9d-925fe107b707-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.548826] env[62183]: DEBUG oslo_concurrency.lockutils [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] Lock "1c180169-ebb5-4870-8e9d-925fe107b707-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.548998] env[62183]: DEBUG nova.compute.manager [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] No waiting events found dispatching network-vif-plugged-78754a06-2366-4beb-bc13-95b404d4acf6 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1033.549188] env[62183]: WARNING nova.compute.manager [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Received unexpected event network-vif-plugged-78754a06-2366-4beb-bc13-95b404d4acf6 for instance with vm_state building and task_state spawning. [ 1033.549358] env[62183]: DEBUG nova.compute.manager [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Received event network-changed-78754a06-2366-4beb-bc13-95b404d4acf6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.549523] env[62183]: DEBUG nova.compute.manager [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Refreshing instance network info cache due to event network-changed-78754a06-2366-4beb-bc13-95b404d4acf6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1033.549701] env[62183]: DEBUG oslo_concurrency.lockutils [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] Acquiring lock "refresh_cache-1c180169-ebb5-4870-8e9d-925fe107b707" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.549840] env[62183]: DEBUG oslo_concurrency.lockutils [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] Acquired lock "refresh_cache-1c180169-ebb5-4870-8e9d-925fe107b707" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.549997] env[62183]: DEBUG nova.network.neutron [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Refreshing network info cache for port 78754a06-2366-4beb-bc13-95b404d4acf6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1033.616155] env[62183]: DEBUG oslo_vmware.api [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387465, 'name': PowerOnVM_Task, 'duration_secs': 0.511628} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.616428] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1033.616634] env[62183]: INFO nova.compute.manager [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Took 7.61 seconds to spawn the instance on the hypervisor. [ 1033.616816] env[62183]: DEBUG nova.compute.manager [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.617618] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e984e8-109d-4cc3-8fa1-d698b07ebfef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.749970] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5200cab2-3344-f3d4-2b1d-9f21c592fd8d, 'name': SearchDatastore_Task, 'duration_secs': 0.012067} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.750424] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.750771] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1033.751128] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.751368] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.751651] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1033.752014] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07d35ccc-3d99-4a57-88fb-99c6e4bbcf1c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.760927] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1033.761204] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1033.762192] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35299347-f158-40b8-ad34-3a73e10fe625 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.767940] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1033.767940] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525f47d9-5698-7238-14c3-6fcdd4d01a43" [ 1033.767940] env[62183]: _type = "Task" [ 1033.767940] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.775257] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525f47d9-5698-7238-14c3-6fcdd4d01a43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.020329] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387468, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.037462] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.037722] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.037939] env[62183]: DEBUG nova.objects.instance [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lazy-loading 'resources' on Instance uuid dbd34888-a333-40b4-ae9c-df5541b16704 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.060255] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5bcee71-3c8d-42d3-8f41-8a00ad2bf7c0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.081902] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9eb3a9-9284-4283-b07e-aec402650bff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.091894] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance 'a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a' progress to 83 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1034.138989] env[62183]: DEBUG nova.compute.manager [req-614bcde2-d774-43bc-b1c2-2a6589e80af1 req-39202b2e-640c-433d-af73-ac7db5abdda1 service nova] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Received event network-vif-deleted-b7e8fcf5-0b5e-4b35-af6b-f18163968d76 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.139559] env[62183]: INFO nova.compute.manager [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Took 15.52 seconds to build instance. [ 1034.277389] env[62183]: DEBUG nova.network.neutron [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Updated VIF entry in instance network info cache for port 78754a06-2366-4beb-bc13-95b404d4acf6. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1034.277789] env[62183]: DEBUG nova.network.neutron [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Updating instance_info_cache with network_info: [{"id": "78754a06-2366-4beb-bc13-95b404d4acf6", "address": "fa:16:3e:5f:09:88", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78754a06-23", "ovs_interfaceid": "78754a06-2366-4beb-bc13-95b404d4acf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.282446] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525f47d9-5698-7238-14c3-6fcdd4d01a43, 'name': SearchDatastore_Task, 'duration_secs': 0.008037} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.283271] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8dd9e38-a6b4-4719-8bc7-46e8b79804fd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.288325] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1034.288325] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]523025de-1040-dd0a-52eb-1dace7b6009a" [ 1034.288325] env[62183]: _type = "Task" [ 1034.288325] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.297976] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]523025de-1040-dd0a-52eb-1dace7b6009a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.520538] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387468, 'name': ReconfigVM_Task, 'duration_secs': 0.920296} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.520838] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86/488d1c9b-fd4f-4b0d-bd05-9f42c2132b86.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.521403] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2fa9f371-e9f1-4159-bc2f-82dede656541 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.527209] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1034.527209] env[62183]: value = "task-1387469" [ 1034.527209] env[62183]: _type = "Task" [ 1034.527209] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.534330] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387469, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.600802] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.601232] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5966056-f6a1-4ee3-bf1d-8c6b3aabda63 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.608864] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1034.608864] env[62183]: value = "task-1387470" [ 1034.608864] env[62183]: _type = "Task" [ 1034.608864] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.616702] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387470, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.641969] env[62183]: DEBUG oslo_concurrency.lockutils [None req-65d60f84-a421-4ef2-b9f2-af4806b5b4bf tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "f51495aa-32b3-429c-8421-65f0a2587ea8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.034s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.674125] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b4b08f-6f48-4c20-b5b4-2ba5f688e830 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.680479] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8505c97-fe23-4451-b7c0-46256acd38bb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.712299] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0b624c-750c-460a-a746-523bc1a65c7d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.719726] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0665ec14-b7a9-49b0-8535-bb0a72165bfd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.733113] env[62183]: DEBUG nova.compute.provider_tree [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.784500] env[62183]: DEBUG oslo_concurrency.lockutils [req-64313879-7e50-4f37-9087-eb2df546242f req-70738c34-f341-4cfc-9a55-44f306fd8e79 service nova] Releasing lock "refresh_cache-1c180169-ebb5-4870-8e9d-925fe107b707" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.801561] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]523025de-1040-dd0a-52eb-1dace7b6009a, 'name': SearchDatastore_Task, 'duration_secs': 0.009425} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.802125] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.802187] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 1c180169-ebb5-4870-8e9d-925fe107b707/1c180169-ebb5-4870-8e9d-925fe107b707.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1034.802449] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7158494-daa5-442b-b47a-f83f4b478b62 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.810066] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1034.810066] env[62183]: value = "task-1387471" [ 1034.810066] env[62183]: _type = "Task" [ 1034.810066] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.816842] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387471, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.039086] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387469, 'name': Rename_Task, 'duration_secs': 0.136755} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.039086] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.039408] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-beafdb7b-3f22-401d-b6fa-46b39524269e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.046929] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1035.046929] env[62183]: value = "task-1387472" [ 1035.046929] env[62183]: _type = "Task" [ 1035.046929] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.055047] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387472, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.120577] env[62183]: DEBUG oslo_vmware.api [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387470, 'name': PowerOnVM_Task, 'duration_secs': 0.340963} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.121017] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1035.121281] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b699a5cd-c8c2-43e0-8828-df0f95733ecc tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance 'a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a' progress to 100 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1035.236512] env[62183]: DEBUG nova.scheduler.client.report [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1035.319218] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387471, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467727} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.319493] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 1c180169-ebb5-4870-8e9d-925fe107b707/1c180169-ebb5-4870-8e9d-925fe107b707.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1035.319760] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.320038] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf6387b5-e61d-4fea-aedc-20e3e16a5aaf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.326022] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1035.326022] env[62183]: value = "task-1387473" [ 1035.326022] env[62183]: _type = "Task" [ 1035.326022] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.333388] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387473, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.399919] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquiring lock "f51495aa-32b3-429c-8421-65f0a2587ea8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.400375] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "f51495aa-32b3-429c-8421-65f0a2587ea8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.400741] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquiring lock "f51495aa-32b3-429c-8421-65f0a2587ea8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.401253] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "f51495aa-32b3-429c-8421-65f0a2587ea8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.401253] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "f51495aa-32b3-429c-8421-65f0a2587ea8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.403621] env[62183]: INFO nova.compute.manager [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Terminating instance [ 1035.405667] env[62183]: DEBUG nova.compute.manager [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.405859] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.406708] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cac6609-a833-45d6-8181-a2f80876e526 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.414915] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.415189] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61acaec0-8ce7-4cf1-a824-b49d93f84a31 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.421460] env[62183]: DEBUG oslo_vmware.api [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1035.421460] env[62183]: value = "task-1387474" [ 1035.421460] env[62183]: _type = "Task" [ 1035.421460] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.430774] env[62183]: DEBUG oslo_vmware.api [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387474, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.556647] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387472, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.741988] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.704s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.761846] env[62183]: INFO nova.scheduler.client.report [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Deleted allocations for instance dbd34888-a333-40b4-ae9c-df5541b16704 [ 1035.835892] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387473, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064044} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.836220] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1035.837042] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e465950e-b568-4751-8383-71443372249e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.859895] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 1c180169-ebb5-4870-8e9d-925fe107b707/1c180169-ebb5-4870-8e9d-925fe107b707.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1035.860219] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e47dab73-a12c-4076-842f-a234ecfecac3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.878996] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1035.878996] env[62183]: value = "task-1387475" [ 1035.878996] env[62183]: _type = "Task" [ 1035.878996] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.887434] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387475, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.931202] env[62183]: DEBUG oslo_vmware.api [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387474, 'name': PowerOffVM_Task, 'duration_secs': 0.474627} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.931500] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1035.931693] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1035.931948] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d2af1df-6595-43a5-ad2d-55ea08688e3b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.018120] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1036.018398] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1036.018597] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Deleting the datastore file [datastore1] f51495aa-32b3-429c-8421-65f0a2587ea8 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.018873] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6769baa-e8bc-4ebc-89d3-1386638ede19 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.026273] env[62183]: DEBUG oslo_vmware.api [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for the task: (returnval){ [ 1036.026273] env[62183]: value = "task-1387477" [ 1036.026273] env[62183]: _type = "Task" [ 1036.026273] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.033956] env[62183]: DEBUG oslo_vmware.api [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387477, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.056050] env[62183]: DEBUG oslo_vmware.api [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387472, 'name': PowerOnVM_Task, 'duration_secs': 0.917908} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.056320] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.056533] env[62183]: INFO nova.compute.manager [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Took 7.76 seconds to spawn the instance on the hypervisor. [ 1036.056712] env[62183]: DEBUG nova.compute.manager [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.057576] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822add4a-6041-40e7-b398-d7cdb27607ff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.269100] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dddab755-c065-48ec-bff1-61019ee0f354 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "dbd34888-a333-40b4-ae9c-df5541b16704" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.059s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.394242] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387475, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.536865] env[62183]: DEBUG oslo_vmware.api [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Task: {'id': task-1387477, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188006} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.537017] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.537111] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.537489] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.537716] env[62183]: INFO nova.compute.manager [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1036.537976] env[62183]: DEBUG oslo.service.loopingcall [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.538217] env[62183]: DEBUG nova.compute.manager [-] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.538410] env[62183]: DEBUG nova.network.neutron [-] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1036.574346] env[62183]: INFO nova.compute.manager [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Took 17.56 seconds to build instance. [ 1036.882487] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "a5018b4e-016d-4c61-bfbf-2c039ab79499" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.882732] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.882963] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "a5018b4e-016d-4c61-bfbf-2c039ab79499-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.883175] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.883356] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.886720] env[62183]: INFO nova.compute.manager [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Terminating instance [ 1036.888702] env[62183]: DEBUG nova.compute.manager [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1036.888914] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1036.893586] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ce8dcc-e711-4001-9340-2f86c06ccabe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.897668] env[62183]: DEBUG nova.compute.manager [req-6e822e5e-9856-4910-8c4d-69458d9746e0 req-c0657744-390f-4e82-92b7-a9b9c286c2a2 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Received event network-vif-deleted-1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1036.897830] env[62183]: INFO nova.compute.manager [req-6e822e5e-9856-4910-8c4d-69458d9746e0 req-c0657744-390f-4e82-92b7-a9b9c286c2a2 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Neutron deleted interface 1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8; detaching it from the instance and deleting it from the info cache [ 1036.898017] env[62183]: DEBUG nova.network.neutron [req-6e822e5e-9856-4910-8c4d-69458d9746e0 req-c0657744-390f-4e82-92b7-a9b9c286c2a2 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.905148] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387475, 'name': ReconfigVM_Task, 'duration_secs': 0.548632} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.907920] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 1c180169-ebb5-4870-8e9d-925fe107b707/1c180169-ebb5-4870-8e9d-925fe107b707.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.908372] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1036.909144] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3b0c17c-932b-41fb-8190-2ffe2218f53d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.910620] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7969250-971e-46f0-9881-887c61a6740d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.917153] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1036.917153] env[62183]: value = "task-1387478" [ 1036.917153] env[62183]: _type = "Task" [ 1036.917153] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.918749] env[62183]: DEBUG oslo_vmware.api [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1036.918749] env[62183]: value = "task-1387479" [ 1036.918749] env[62183]: _type = "Task" [ 1036.918749] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.929727] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387478, 'name': Rename_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.933049] env[62183]: DEBUG oslo_vmware.api [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387479, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.079410] env[62183]: DEBUG oslo_concurrency.lockutils [None req-dc39afa8-5c96-454c-9187-4ae0570db95f tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.069s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.215107] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.215493] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.215765] env[62183]: DEBUG nova.compute.manager [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Going to confirm migration 5 {{(pid=62183) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1037.331461] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.331769] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.332038] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.332286] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.333043] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.335360] env[62183]: INFO nova.compute.manager [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Terminating instance [ 1037.337496] env[62183]: DEBUG nova.compute.manager [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1037.337695] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.338599] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c7731b-9b7b-47c4-bde7-dbd2af837d5e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.345861] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.346124] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4db06e38-e8a3-431a-bfbb-77c04bb198b3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.352385] env[62183]: DEBUG oslo_vmware.api [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1037.352385] env[62183]: value = "task-1387480" [ 1037.352385] env[62183]: _type = "Task" [ 1037.352385] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.359838] env[62183]: DEBUG oslo_vmware.api [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.367236] env[62183]: DEBUG nova.network.neutron [-] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.400566] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3672ae53-1065-4722-925a-c67bc8807103 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.409341] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c98f1b-521c-46b7-97c0-dfc7c83197c5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.430936] env[62183]: DEBUG oslo_vmware.api [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387479, 'name': PowerOffVM_Task, 'duration_secs': 0.178036} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.433865] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.434066] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.434338] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387478, 'name': Rename_Task, 'duration_secs': 0.159482} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.441785] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e07c6dc9-8cf2-4d86-9dac-94568c3b28e3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.443633] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.443975] env[62183]: DEBUG nova.compute.manager [req-6e822e5e-9856-4910-8c4d-69458d9746e0 req-c0657744-390f-4e82-92b7-a9b9c286c2a2 service nova] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Detach interface failed, port_id=1afd9c1a-eae0-4fe5-ad5b-ad363c2550e8, reason: Instance f51495aa-32b3-429c-8421-65f0a2587ea8 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1037.444441] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9680abc-9138-4510-83eb-513b5b99b6ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.451584] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1037.451584] env[62183]: value = "task-1387482" [ 1037.451584] env[62183]: _type = "Task" [ 1037.451584] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.460131] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387482, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.513236] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.513533] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.513729] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Deleting the datastore file [datastore1] a5018b4e-016d-4c61-bfbf-2c039ab79499 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.514042] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c975ece7-6522-411f-847f-0a5090dd0548 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.521988] env[62183]: DEBUG oslo_vmware.api [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for the task: (returnval){ [ 1037.521988] env[62183]: value = "task-1387483" [ 1037.521988] env[62183]: _type = "Task" [ 1037.521988] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.534086] env[62183]: DEBUG oslo_vmware.api [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387483, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.767760] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.768025] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.768145] env[62183]: DEBUG nova.network.neutron [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1037.768344] env[62183]: DEBUG nova.objects.instance [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'info_cache' on Instance uuid a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.865420] env[62183]: DEBUG oslo_vmware.api [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387480, 'name': PowerOffVM_Task, 'duration_secs': 0.276861} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.865666] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.865839] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.870647] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b04245c6-c28d-43f3-b939-5bc76f7aabf8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.872505] env[62183]: INFO nova.compute.manager [-] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Took 1.33 seconds to deallocate network for instance. [ 1037.952735] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.952983] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.953216] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleting the datastore file [datastore1] 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.960894] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63bdddc8-da2d-4143-9c8e-d06316eb9311 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.968322] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387482, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.969666] env[62183]: DEBUG oslo_vmware.api [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1037.969666] env[62183]: value = "task-1387485" [ 1037.969666] env[62183]: _type = "Task" [ 1037.969666] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.977465] env[62183]: DEBUG oslo_vmware.api [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387485, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.032550] env[62183]: DEBUG oslo_vmware.api [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Task: {'id': task-1387483, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222362} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.032991] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.033318] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.033617] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.033848] env[62183]: INFO nova.compute.manager [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1038.034092] env[62183]: DEBUG oslo.service.loopingcall [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.034430] env[62183]: DEBUG nova.compute.manager [-] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1038.034594] env[62183]: DEBUG nova.network.neutron [-] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1038.384771] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.385097] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.385362] env[62183]: DEBUG nova.objects.instance [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lazy-loading 'resources' on Instance uuid f51495aa-32b3-429c-8421-65f0a2587ea8 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.461586] env[62183]: DEBUG oslo_vmware.api [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387482, 'name': PowerOnVM_Task, 'duration_secs': 0.705922} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.461802] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.462009] env[62183]: INFO nova.compute.manager [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Took 7.93 seconds to spawn the instance on the hypervisor. [ 1038.462195] env[62183]: DEBUG nova.compute.manager [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.462957] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6fcff45-a70a-453e-ac60-c57425cf996a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.478095] env[62183]: DEBUG oslo_vmware.api [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387485, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.358758} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.478331] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.478515] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.478695] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.478864] env[62183]: INFO nova.compute.manager [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1038.479112] env[62183]: DEBUG oslo.service.loopingcall [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.479305] env[62183]: DEBUG nova.compute.manager [-] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1038.479401] env[62183]: DEBUG nova.network.neutron [-] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1038.789925] env[62183]: DEBUG nova.network.neutron [-] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.919148] env[62183]: DEBUG nova.compute.manager [req-444a977d-6985-44c5-871a-387926ffbc64 req-d7ef7502-dff3-4d4c-b2ab-0c7ed1406267 service nova] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Received event network-vif-deleted-f48d3c73-edab-4cf3-94cc-498503b5c465 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.919403] env[62183]: DEBUG nova.compute.manager [req-444a977d-6985-44c5-871a-387926ffbc64 req-d7ef7502-dff3-4d4c-b2ab-0c7ed1406267 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Received event network-vif-deleted-f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.919543] env[62183]: INFO nova.compute.manager [req-444a977d-6985-44c5-871a-387926ffbc64 req-d7ef7502-dff3-4d4c-b2ab-0c7ed1406267 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Neutron deleted interface f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75; detaching it from the instance and deleting it from the info cache [ 1038.919713] env[62183]: DEBUG nova.network.neutron [req-444a977d-6985-44c5-871a-387926ffbc64 req-d7ef7502-dff3-4d4c-b2ab-0c7ed1406267 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.981589] env[62183]: INFO nova.compute.manager [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Took 16.08 seconds to build instance. [ 1038.998792] env[62183]: DEBUG nova.network.neutron [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [{"id": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "address": "fa:16:3e:49:83:73", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbba3a12a-3b", "ovs_interfaceid": "bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.016830] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6a5199-cb3d-4bf5-af87-e50d21bcf6b2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.024434] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162f244b-f79d-4e30-9481-91c190408870 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.055513] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8003f6-8f7b-4b81-b39f-2f1c7d17800e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.062988] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498c0696-cc6c-4f7a-93a7-4b1d814b8903 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.078278] env[62183]: DEBUG nova.compute.provider_tree [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.202168] env[62183]: DEBUG nova.network.neutron [-] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.292815] env[62183]: INFO nova.compute.manager [-] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Took 1.26 seconds to deallocate network for instance. [ 1039.426195] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a804a24f-dcab-428e-a0bf-0649e19a2b60 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.435664] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d0ab5c-80a8-4fa0-aac3-cb2969822a3c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.461928] env[62183]: DEBUG nova.compute.manager [req-444a977d-6985-44c5-871a-387926ffbc64 req-d7ef7502-dff3-4d4c-b2ab-0c7ed1406267 service nova] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Detach interface failed, port_id=f0b525f4-3eda-4ba7-ac7e-bbc5cabf7e75, reason: Instance 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1039.485290] env[62183]: DEBUG oslo_concurrency.lockutils [None req-3f4c7d9b-d828-49a3-908b-04ca48dfb6c0 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.589s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.500923] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.501216] env[62183]: DEBUG nova.objects.instance [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'migration_context' on Instance uuid a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.581153] env[62183]: DEBUG nova.scheduler.client.report [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1039.643224] env[62183]: DEBUG nova.compute.manager [req-4459636c-6114-45bd-bf90-082a6e7a68e5 req-295ea581-d6ec-4b28-a616-d3045b0cf66e service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Received event network-changed-78754a06-2366-4beb-bc13-95b404d4acf6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.643431] env[62183]: DEBUG nova.compute.manager [req-4459636c-6114-45bd-bf90-082a6e7a68e5 req-295ea581-d6ec-4b28-a616-d3045b0cf66e service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Refreshing instance network info cache due to event network-changed-78754a06-2366-4beb-bc13-95b404d4acf6. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.643654] env[62183]: DEBUG oslo_concurrency.lockutils [req-4459636c-6114-45bd-bf90-082a6e7a68e5 req-295ea581-d6ec-4b28-a616-d3045b0cf66e service nova] Acquiring lock "refresh_cache-1c180169-ebb5-4870-8e9d-925fe107b707" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.643802] env[62183]: DEBUG oslo_concurrency.lockutils [req-4459636c-6114-45bd-bf90-082a6e7a68e5 req-295ea581-d6ec-4b28-a616-d3045b0cf66e service nova] Acquired lock "refresh_cache-1c180169-ebb5-4870-8e9d-925fe107b707" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.643965] env[62183]: DEBUG nova.network.neutron [req-4459636c-6114-45bd-bf90-082a6e7a68e5 req-295ea581-d6ec-4b28-a616-d3045b0cf66e service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Refreshing network info cache for port 78754a06-2366-4beb-bc13-95b404d4acf6 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1039.706891] env[62183]: INFO nova.compute.manager [-] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Took 1.23 seconds to deallocate network for instance. [ 1039.799116] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.004275] env[62183]: DEBUG nova.objects.base [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1040.005285] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6b01b8-b42b-47c4-bc4e-4d4cd7e43961 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.026994] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-547f503a-f9f3-4f54-9114-5de421b0e13c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.031946] env[62183]: DEBUG oslo_vmware.api [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1040.031946] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52604e2d-355f-efe0-8246-02d36302ef39" [ 1040.031946] env[62183]: _type = "Task" [ 1040.031946] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.039214] env[62183]: DEBUG oslo_vmware.api [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52604e2d-355f-efe0-8246-02d36302ef39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.086093] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.701s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.088327] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.289s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.088577] env[62183]: DEBUG nova.objects.instance [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lazy-loading 'resources' on Instance uuid a5018b4e-016d-4c61-bfbf-2c039ab79499 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.109820] env[62183]: INFO nova.scheduler.client.report [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Deleted allocations for instance f51495aa-32b3-429c-8421-65f0a2587ea8 [ 1040.213990] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.392598] env[62183]: DEBUG nova.network.neutron [req-4459636c-6114-45bd-bf90-082a6e7a68e5 req-295ea581-d6ec-4b28-a616-d3045b0cf66e service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Updated VIF entry in instance network info cache for port 78754a06-2366-4beb-bc13-95b404d4acf6. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1040.393054] env[62183]: DEBUG nova.network.neutron [req-4459636c-6114-45bd-bf90-082a6e7a68e5 req-295ea581-d6ec-4b28-a616-d3045b0cf66e service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Updating instance_info_cache with network_info: [{"id": "78754a06-2366-4beb-bc13-95b404d4acf6", "address": "fa:16:3e:5f:09:88", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78754a06-23", "ovs_interfaceid": "78754a06-2366-4beb-bc13-95b404d4acf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.543016] env[62183]: DEBUG oslo_vmware.api [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52604e2d-355f-efe0-8246-02d36302ef39, 'name': SearchDatastore_Task, 'duration_secs': 0.007011} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.543280] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.617537] env[62183]: DEBUG oslo_concurrency.lockutils [None req-a4f77eb0-fa02-458e-a4fa-98be4d5721e0 tempest-ServersNegativeTestMultiTenantJSON-566463170 tempest-ServersNegativeTestMultiTenantJSON-566463170-project-member] Lock "f51495aa-32b3-429c-8421-65f0a2587ea8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.217s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.706142] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25dec5ac-0d77-44aa-bd4c-d5c0651c678e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.714647] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4425c6f-85ac-4ffc-a293-5ccedfede038 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.744855] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec508c19-1f15-4a4a-a107-d144f7088951 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.752790] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383dfce9-891f-46a6-be7e-d8e30cedbde3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.768424] env[62183]: DEBUG nova.compute.provider_tree [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.895723] env[62183]: DEBUG oslo_concurrency.lockutils [req-4459636c-6114-45bd-bf90-082a6e7a68e5 req-295ea581-d6ec-4b28-a616-d3045b0cf66e service nova] Releasing lock "refresh_cache-1c180169-ebb5-4870-8e9d-925fe107b707" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.272200] env[62183]: DEBUG nova.scheduler.client.report [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1041.778163] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.690s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.780428] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.567s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.780669] env[62183]: DEBUG nova.objects.instance [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lazy-loading 'resources' on Instance uuid 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.799444] env[62183]: INFO nova.scheduler.client.report [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Deleted allocations for instance a5018b4e-016d-4c61-bfbf-2c039ab79499 [ 1042.312988] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d5db957a-efd7-4e8b-9697-2cedfe1b49c3 tempest-ListServerFiltersTestJSON-1267328345 tempest-ListServerFiltersTestJSON-1267328345-project-member] Lock "a5018b4e-016d-4c61-bfbf-2c039ab79499" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.430s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.413172] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f37d357-2d98-4561-9edc-fc9acffafd5e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.422344] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9442a0b-c455-4cdf-9ab8-618325b900db {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.455421] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d455149-0def-4410-afd2-40d47ec602a0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.465553] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ab3eb8-d72e-4926-b9cc-f1ac5b696b22 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.482402] env[62183]: DEBUG nova.compute.provider_tree [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.985433] env[62183]: DEBUG nova.scheduler.client.report [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1043.491208] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.494061] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.950s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.518038] env[62183]: INFO nova.scheduler.client.report [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted allocations for instance 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86 [ 1044.029134] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0816a909-9e33-4ab6-b942-9de5724dde9b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "488d1c9b-fd4f-4b0d-bd05-9f42c2132b86" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.697s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.078616] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a64cb4-4c78-4a97-b6f6-4e16c094da71 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.086633] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db0f08f-553d-4b5a-ac54-91a8bde75e89 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.116375] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a7ba5b-7608-4513-84b9-290524e9041f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.123350] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be0fb6c-d7b5-4d59-bd10-e6b8462e6945 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.136163] env[62183]: DEBUG nova.compute.provider_tree [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.639321] env[62183]: DEBUG nova.scheduler.client.report [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1045.583198] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "78db4224-11cb-468c-9c26-d470cac12b0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.583435] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "78db4224-11cb-468c-9c26-d470cac12b0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.649628] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.156s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.086147] env[62183]: DEBUG nova.compute.manager [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1046.206992] env[62183]: INFO nova.scheduler.client.report [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted allocation for migration e4b22d4f-df7b-4efe-a48e-ddc16cb72b96 [ 1046.604210] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.604498] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.606042] env[62183]: INFO nova.compute.claims [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1046.690818] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.714014] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6d8a25f9-fe78-4c0c-a63d-68d21d3e6231 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.498s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.715054] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.024s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.715333] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.715553] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.715733] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.717566] env[62183]: INFO nova.compute.manager [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Terminating instance [ 1046.719246] env[62183]: DEBUG nova.compute.manager [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1046.719447] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1046.720314] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0b4db6-5033-44e5-b259-c7be8d25e8e4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.727890] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1046.728152] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7e64cde-ac57-48e6-9196-ce2066ff9c9d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.735335] env[62183]: DEBUG oslo_vmware.api [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1046.735335] env[62183]: value = "task-1387486" [ 1046.735335] env[62183]: _type = "Task" [ 1046.735335] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.742764] env[62183]: DEBUG oslo_vmware.api [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387486, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.245246] env[62183]: DEBUG oslo_vmware.api [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387486, 'name': PowerOffVM_Task, 'duration_secs': 0.228443} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.245536] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1047.245712] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1047.245968] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5306dd62-727a-47d6-9e5c-ceb01fce05dc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.314037] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1047.314202] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1047.314350] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleting the datastore file [datastore2] a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1047.314639] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b29c9bdf-f2f5-43ff-b00f-0e2287ac7ac6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.323827] env[62183]: DEBUG oslo_vmware.api [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1047.323827] env[62183]: value = "task-1387488" [ 1047.323827] env[62183]: _type = "Task" [ 1047.323827] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.331194] env[62183]: DEBUG oslo_vmware.api [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387488, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.654109] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1047.654385] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1047.654434] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1047.695407] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451f7f05-2f35-4e10-86bc-6ceb45d9d82c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.703513] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fdd7dbd-bbf8-457b-88ee-4e74b63b997b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.732947] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889623d6-c7aa-4e43-9e30-4393d465f52b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.739920] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31a10b9-5e68-426d-af32-2a40398b4cdd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.752562] env[62183]: DEBUG nova.compute.provider_tree [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.832498] env[62183]: DEBUG oslo_vmware.api [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387488, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166146} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.832755] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1047.832944] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1047.833178] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1047.833348] env[62183]: INFO nova.compute.manager [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1047.833593] env[62183]: DEBUG oslo.service.loopingcall [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.833784] env[62183]: DEBUG nova.compute.manager [-] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1047.834131] env[62183]: DEBUG nova.network.neutron [-] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1048.253094] env[62183]: DEBUG nova.compute.manager [req-20e8ffbb-6a56-4a69-8629-b6e551345911 req-ab00ed89-9521-49a8-a14d-8c7a464b7496 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Received event network-vif-deleted-bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1048.253094] env[62183]: INFO nova.compute.manager [req-20e8ffbb-6a56-4a69-8629-b6e551345911 req-ab00ed89-9521-49a8-a14d-8c7a464b7496 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Neutron deleted interface bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e; detaching it from the instance and deleting it from the info cache [ 1048.253094] env[62183]: DEBUG nova.network.neutron [req-20e8ffbb-6a56-4a69-8629-b6e551345911 req-ab00ed89-9521-49a8-a14d-8c7a464b7496 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.256507] env[62183]: DEBUG nova.scheduler.client.report [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1048.731936] env[62183]: DEBUG nova.network.neutron [-] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.756455] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4d2215b7-c95a-4ebe-9c4c-9de812cd72dc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.760746] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.156s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.761250] env[62183]: DEBUG nova.compute.manager [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1048.767650] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017def0b-795f-4fd4-b3d3-7020e9935d02 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.793720] env[62183]: DEBUG nova.compute.manager [req-20e8ffbb-6a56-4a69-8629-b6e551345911 req-ab00ed89-9521-49a8-a14d-8c7a464b7496 service nova] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Detach interface failed, port_id=bba3a12a-3b45-49e4-9cd9-8d5f3a46a81e, reason: Instance a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1049.164508] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Didn't find any instances for network info cache update. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1049.164751] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.164913] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.165112] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.165292] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.165453] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.165598] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.165730] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1049.165880] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.235018] env[62183]: INFO nova.compute.manager [-] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Took 1.40 seconds to deallocate network for instance. [ 1049.266569] env[62183]: DEBUG nova.compute.utils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1049.267914] env[62183]: DEBUG nova.compute.manager [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1049.268099] env[62183]: DEBUG nova.network.neutron [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1049.306398] env[62183]: DEBUG nova.policy [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88efcb2057e84413806f9ed809c260e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eda7e297c551407295b3c67fa0f1dbc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1049.565690] env[62183]: DEBUG nova.network.neutron [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Successfully created port: 6d72dc33-f503-4d8e-b8a9-1ea5979dbe78 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1049.669353] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.669645] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.669824] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.669978] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1049.670877] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcc4b9e-9110-4f3a-ad8f-17568ad9ffad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.678851] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ac3f2f-13a5-4e0d-8f25-e0857b9eaca8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.692471] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c907986-62a9-460f-84a1-d2fb944aaa7b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.698468] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc25f23-02ba-469e-b9ab-4f8988d3ab70 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.727136] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180385MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1049.727317] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.727519] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.741602] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.771671] env[62183]: DEBUG nova.compute.manager [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1050.756191] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1050.756447] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 8c4b68d6-fedd-408f-a449-aace7400014a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1050.756518] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 88846abe-9e6e-4a88-bf6e-9978610213da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1050.756737] env[62183]: WARNING nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1050.756886] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 1c180169-ebb5-4870-8e9d-925fe107b707 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1050.757061] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 78db4224-11cb-468c-9c26-d470cac12b0e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1050.757306] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1050.757463] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1050.784267] env[62183]: DEBUG nova.compute.manager [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1050.811653] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1050.812018] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1050.812308] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1050.812532] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1050.812687] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1050.812840] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1050.813084] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1050.813256] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1050.813433] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1050.813599] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1050.813785] env[62183]: DEBUG nova.virt.hardware [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1050.815270] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805bb57a-7053-4610-8007-94e680abff46 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.826925] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f69eace-a4eb-41a4-bdfa-44195ec287c4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.851400] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10addcaa-e3f3-4faa-965b-d362901e45d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.858115] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ff03d1-81b9-4b99-8d35-d75fc854fd4a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.888587] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9dba25b-108f-4bee-9411-5336d83d11af {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.896007] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c5d292-8cb6-4f84-83a2-b053bc44b836 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.909044] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.931905] env[62183]: DEBUG nova.compute.manager [req-a36fd2fd-d118-41b0-9d8b-bd7c5c88f3b0 req-c2082c43-9eab-47e6-a212-e8b87242c961 service nova] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Received event network-vif-plugged-6d72dc33-f503-4d8e-b8a9-1ea5979dbe78 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1050.932144] env[62183]: DEBUG oslo_concurrency.lockutils [req-a36fd2fd-d118-41b0-9d8b-bd7c5c88f3b0 req-c2082c43-9eab-47e6-a212-e8b87242c961 service nova] Acquiring lock "78db4224-11cb-468c-9c26-d470cac12b0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.932359] env[62183]: DEBUG oslo_concurrency.lockutils [req-a36fd2fd-d118-41b0-9d8b-bd7c5c88f3b0 req-c2082c43-9eab-47e6-a212-e8b87242c961 service nova] Lock "78db4224-11cb-468c-9c26-d470cac12b0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.932543] env[62183]: DEBUG oslo_concurrency.lockutils [req-a36fd2fd-d118-41b0-9d8b-bd7c5c88f3b0 req-c2082c43-9eab-47e6-a212-e8b87242c961 service nova] Lock "78db4224-11cb-468c-9c26-d470cac12b0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.932748] env[62183]: DEBUG nova.compute.manager [req-a36fd2fd-d118-41b0-9d8b-bd7c5c88f3b0 req-c2082c43-9eab-47e6-a212-e8b87242c961 service nova] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] No waiting events found dispatching network-vif-plugged-6d72dc33-f503-4d8e-b8a9-1ea5979dbe78 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1050.932862] env[62183]: WARNING nova.compute.manager [req-a36fd2fd-d118-41b0-9d8b-bd7c5c88f3b0 req-c2082c43-9eab-47e6-a212-e8b87242c961 service nova] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Received unexpected event network-vif-plugged-6d72dc33-f503-4d8e-b8a9-1ea5979dbe78 for instance with vm_state building and task_state spawning. [ 1051.021500] env[62183]: DEBUG nova.network.neutron [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Successfully updated port: 6d72dc33-f503-4d8e-b8a9-1ea5979dbe78 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1051.411714] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1051.523743] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "refresh_cache-78db4224-11cb-468c-9c26-d470cac12b0e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.523906] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "refresh_cache-78db4224-11cb-468c-9c26-d470cac12b0e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.524081] env[62183]: DEBUG nova.network.neutron [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1051.916333] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1051.916662] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.189s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.916871] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.175s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.917100] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.935743] env[62183]: INFO nova.scheduler.client.report [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted allocations for instance a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a [ 1052.061339] env[62183]: DEBUG nova.network.neutron [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1052.191443] env[62183]: DEBUG nova.network.neutron [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Updating instance_info_cache with network_info: [{"id": "6d72dc33-f503-4d8e-b8a9-1ea5979dbe78", "address": "fa:16:3e:40:00:5b", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d72dc33-f5", "ovs_interfaceid": "6d72dc33-f503-4d8e-b8a9-1ea5979dbe78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.444261] env[62183]: DEBUG oslo_concurrency.lockutils [None req-2caeeb8f-ab3c-4459-9f38-ed2c2697f9e1 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.729s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.694564] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "refresh_cache-78db4224-11cb-468c-9c26-d470cac12b0e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.694926] env[62183]: DEBUG nova.compute.manager [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Instance network_info: |[{"id": "6d72dc33-f503-4d8e-b8a9-1ea5979dbe78", "address": "fa:16:3e:40:00:5b", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d72dc33-f5", "ovs_interfaceid": "6d72dc33-f503-4d8e-b8a9-1ea5979dbe78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1052.695478] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:00:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15ff34f9-4b02-4be1-b433-3ec4bd1b37c2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d72dc33-f503-4d8e-b8a9-1ea5979dbe78', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1052.702965] env[62183]: DEBUG oslo.service.loopingcall [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.703929] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1052.704019] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6fa1295-61d6-4eb4-9251-d13b45733de2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.725719] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1052.725719] env[62183]: value = "task-1387489" [ 1052.725719] env[62183]: _type = "Task" [ 1052.725719] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.733855] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387489, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.959123] env[62183]: DEBUG nova.compute.manager [req-34267c5e-3be2-4386-aec8-bff47d11d03a req-0b51f2f9-2734-4492-a48f-0a3914f7f8b1 service nova] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Received event network-changed-6d72dc33-f503-4d8e-b8a9-1ea5979dbe78 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.959360] env[62183]: DEBUG nova.compute.manager [req-34267c5e-3be2-4386-aec8-bff47d11d03a req-0b51f2f9-2734-4492-a48f-0a3914f7f8b1 service nova] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Refreshing instance network info cache due to event network-changed-6d72dc33-f503-4d8e-b8a9-1ea5979dbe78. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1052.959632] env[62183]: DEBUG oslo_concurrency.lockutils [req-34267c5e-3be2-4386-aec8-bff47d11d03a req-0b51f2f9-2734-4492-a48f-0a3914f7f8b1 service nova] Acquiring lock "refresh_cache-78db4224-11cb-468c-9c26-d470cac12b0e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.959854] env[62183]: DEBUG oslo_concurrency.lockutils [req-34267c5e-3be2-4386-aec8-bff47d11d03a req-0b51f2f9-2734-4492-a48f-0a3914f7f8b1 service nova] Acquired lock "refresh_cache-78db4224-11cb-468c-9c26-d470cac12b0e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.959961] env[62183]: DEBUG nova.network.neutron [req-34267c5e-3be2-4386-aec8-bff47d11d03a req-0b51f2f9-2734-4492-a48f-0a3914f7f8b1 service nova] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Refreshing network info cache for port 6d72dc33-f503-4d8e-b8a9-1ea5979dbe78 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1053.235748] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387489, 'name': CreateVM_Task, 'duration_secs': 0.308936} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.235967] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1053.236602] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.236791] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.237129] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1053.237409] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e80ee1da-4a12-4e51-b236-fd1a8f1cf2c4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.241765] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1053.241765] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]527f6d3a-0ee3-7fc6-2f9c-f1a9d85a7200" [ 1053.241765] env[62183]: _type = "Task" [ 1053.241765] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.248923] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527f6d3a-0ee3-7fc6-2f9c-f1a9d85a7200, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.658815] env[62183]: DEBUG nova.network.neutron [req-34267c5e-3be2-4386-aec8-bff47d11d03a req-0b51f2f9-2734-4492-a48f-0a3914f7f8b1 service nova] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Updated VIF entry in instance network info cache for port 6d72dc33-f503-4d8e-b8a9-1ea5979dbe78. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1053.659184] env[62183]: DEBUG nova.network.neutron [req-34267c5e-3be2-4386-aec8-bff47d11d03a req-0b51f2f9-2734-4492-a48f-0a3914f7f8b1 service nova] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Updating instance_info_cache with network_info: [{"id": "6d72dc33-f503-4d8e-b8a9-1ea5979dbe78", "address": "fa:16:3e:40:00:5b", "network": {"id": "eaebf7a5-e82d-4c17-a378-cdfffda70474", "bridge": "br-int", "label": "tempest-ServersTestJSON-773676908-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eda7e297c551407295b3c67fa0f1dbc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15ff34f9-4b02-4be1-b433-3ec4bd1b37c2", "external-id": "nsx-vlan-transportzone-51", "segmentation_id": 51, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d72dc33-f5", "ovs_interfaceid": "6d72dc33-f503-4d8e-b8a9-1ea5979dbe78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.715409] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "5aed3b2c-627c-4f33-8f22-950e758f822d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.715647] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.751615] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]527f6d3a-0ee3-7fc6-2f9c-f1a9d85a7200, 'name': SearchDatastore_Task, 'duration_secs': 0.012059} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.751911] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.752169] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1053.752441] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.752610] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.752794] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1053.753060] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ca236dd-d575-4f28-9bea-fe513ae36b5b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.760577] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1053.760758] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1053.761453] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41d0d7b3-2b34-435b-9126-93c258eb56a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.766261] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1053.766261] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52c23c2c-6ac8-9503-5c2b-676a6020ff29" [ 1053.766261] env[62183]: _type = "Task" [ 1053.766261] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.773281] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c23c2c-6ac8-9503-5c2b-676a6020ff29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.162329] env[62183]: DEBUG oslo_concurrency.lockutils [req-34267c5e-3be2-4386-aec8-bff47d11d03a req-0b51f2f9-2734-4492-a48f-0a3914f7f8b1 service nova] Releasing lock "refresh_cache-78db4224-11cb-468c-9c26-d470cac12b0e" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.218783] env[62183]: DEBUG nova.compute.manager [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1054.276651] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c23c2c-6ac8-9503-5c2b-676a6020ff29, 'name': SearchDatastore_Task, 'duration_secs': 0.007979} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.277440] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1af75130-5f17-45a7-a78d-2735583f3dc8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.282476] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1054.282476] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52bf3a61-cc7f-f3fa-ccc7-be5cd00c0027" [ 1054.282476] env[62183]: _type = "Task" [ 1054.282476] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.289530] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bf3a61-cc7f-f3fa-ccc7-be5cd00c0027, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.741289] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.741560] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.743081] env[62183]: INFO nova.compute.claims [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.794258] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bf3a61-cc7f-f3fa-ccc7-be5cd00c0027, 'name': SearchDatastore_Task, 'duration_secs': 0.011137} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.794539] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.794796] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 78db4224-11cb-468c-9c26-d470cac12b0e/78db4224-11cb-468c-9c26-d470cac12b0e.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1054.795064] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a189d896-a054-479a-9718-559fb2f94f4b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.802409] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1054.802409] env[62183]: value = "task-1387490" [ 1054.802409] env[62183]: _type = "Task" [ 1054.802409] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.809709] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.313054] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454652} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.313054] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 78db4224-11cb-468c-9c26-d470cac12b0e/78db4224-11cb-468c-9c26-d470cac12b0e.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1055.313054] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1055.313661] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a885aa11-8220-4262-9c29-e1acf5c28c33 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.319378] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1055.319378] env[62183]: value = "task-1387491" [ 1055.319378] env[62183]: _type = "Task" [ 1055.319378] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.326863] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387491, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.821516] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68eda2a3-f1f2-4bdd-bbe6-fc01e1e40d54 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.834064] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf64b5c-bb9b-47e6-ae5a-63e444754915 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.836952] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387491, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068022} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.837257] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1055.838266] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451b9990-966a-4b72-b2c1-291b27e41850 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.864252] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b575288b-0a05-4597-8297-f4365d044655 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.884105] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 78db4224-11cb-468c-9c26-d470cac12b0e/78db4224-11cb-468c-9c26-d470cac12b0e.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1055.884655] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa311333-8c15-4bb9-94c8-88e5ff8109c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.901429] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350aebec-d5e7-41f2-b58e-086b9a1c8048 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.905808] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1055.905808] env[62183]: value = "task-1387492" [ 1055.905808] env[62183]: _type = "Task" [ 1055.905808] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.917242] env[62183]: DEBUG nova.compute.provider_tree [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.923305] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387492, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.416139] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387492, 'name': ReconfigVM_Task, 'duration_secs': 0.309243} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.416497] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 78db4224-11cb-468c-9c26-d470cac12b0e/78db4224-11cb-468c-9c26-d470cac12b0e.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1056.416975] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd9d0414-2594-43b3-832e-fd41d46b1f5d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.420355] env[62183]: DEBUG nova.scheduler.client.report [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1056.424264] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1056.424264] env[62183]: value = "task-1387493" [ 1056.424264] env[62183]: _type = "Task" [ 1056.424264] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.431990] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387493, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.925596] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.184s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.926096] env[62183]: DEBUG nova.compute.manager [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1056.938262] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387493, 'name': Rename_Task, 'duration_secs': 0.132938} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.938528] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1056.938779] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a9040cee-4d32-498f-a174-557da1f5d48e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.945190] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1056.945190] env[62183]: value = "task-1387494" [ 1056.945190] env[62183]: _type = "Task" [ 1056.945190] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.952763] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387494, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.434852] env[62183]: DEBUG nova.compute.utils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1057.436731] env[62183]: DEBUG nova.compute.manager [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1057.436953] env[62183]: DEBUG nova.network.neutron [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1057.454543] env[62183]: DEBUG oslo_vmware.api [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387494, 'name': PowerOnVM_Task, 'duration_secs': 0.438612} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.454782] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1057.454981] env[62183]: INFO nova.compute.manager [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Took 6.67 seconds to spawn the instance on the hypervisor. [ 1057.455220] env[62183]: DEBUG nova.compute.manager [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.456061] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119ea123-e8ef-4db9-a62b-526cfe2d28cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.476255] env[62183]: DEBUG nova.policy [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b85b5dce9bf44da6afd11f100a121f03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '34581a9b5c1943eabc13bbb300a0f086', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1057.729779] env[62183]: DEBUG nova.network.neutron [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Successfully created port: d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1057.940286] env[62183]: DEBUG nova.compute.manager [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1057.973662] env[62183]: INFO nova.compute.manager [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Took 11.38 seconds to build instance. [ 1058.296353] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "8c4b68d6-fedd-408f-a449-aace7400014a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.296644] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.475787] env[62183]: DEBUG oslo_concurrency.lockutils [None req-670931d9-e143-43fb-acd0-0c801e14f2fb tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "78db4224-11cb-468c-9c26-d470cac12b0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.892s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.799422] env[62183]: INFO nova.compute.manager [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Detaching volume 786b9d04-6ed3-4cb1-a773-6db867ff2adb [ 1058.829038] env[62183]: INFO nova.virt.block_device [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Attempting to driver detach volume 786b9d04-6ed3-4cb1-a773-6db867ff2adb from mountpoint /dev/sdb [ 1058.829296] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Volume detach. Driver type: vmdk {{(pid=62183) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1058.829487] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294514', 'volume_id': '786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'name': 'volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c4b68d6-fedd-408f-a449-aace7400014a', 'attached_at': '', 'detached_at': '', 'volume_id': '786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'serial': '786b9d04-6ed3-4cb1-a773-6db867ff2adb'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1058.830422] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae6b78e-485e-40a7-b904-5d0510e5721f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.851536] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39443ee7-5eef-4f71-ae2c-3945d7eca741 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.858327] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2874dae7-49a5-41a6-8db8-9b000bbcdd01 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.877507] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94042541-ea6f-45fd-aa93-a60c777d3faf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.891234] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] The volume has not been displaced from its original location: [datastore1] volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb/volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb.vmdk. No consolidation needed. {{(pid=62183) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1058.896358] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Reconfiguring VM instance instance-0000005a to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1058.896631] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50964ed2-289e-4e34-8cf6-0ea30d743b5f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.913111] env[62183]: DEBUG oslo_vmware.api [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1058.913111] env[62183]: value = "task-1387495" [ 1058.913111] env[62183]: _type = "Task" [ 1058.913111] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.921658] env[62183]: DEBUG oslo_vmware.api [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387495, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.944389] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "78db4224-11cb-468c-9c26-d470cac12b0e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.944657] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "78db4224-11cb-468c-9c26-d470cac12b0e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.944901] env[62183]: DEBUG nova.compute.manager [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1058.945764] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b5ceb8-4baa-464f-9ac5-0af5c6bad562 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.949258] env[62183]: DEBUG nova.compute.manager [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1058.954416] env[62183]: DEBUG nova.compute.manager [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62183) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1058.954806] env[62183]: DEBUG nova.objects.instance [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lazy-loading 'flavor' on Instance uuid 78db4224-11cb-468c-9c26-d470cac12b0e {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.975222] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1058.975514] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1058.975699] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.975889] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1058.976104] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.976335] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1058.976576] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1058.976743] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1058.976914] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1058.977091] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1058.977301] env[62183]: DEBUG nova.virt.hardware [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.978337] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f56fc92-87fc-473e-9150-87ba02457519 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.985197] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a795518a-593f-4434-b25b-ec3185042482 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.112193] env[62183]: DEBUG nova.compute.manager [req-37bb3801-75de-4485-8b7b-82c44bfa4c50 req-a8bf109f-2de2-410e-93bd-22f2878bd4d5 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Received event network-vif-plugged-d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.112429] env[62183]: DEBUG oslo_concurrency.lockutils [req-37bb3801-75de-4485-8b7b-82c44bfa4c50 req-a8bf109f-2de2-410e-93bd-22f2878bd4d5 service nova] Acquiring lock "5aed3b2c-627c-4f33-8f22-950e758f822d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.112645] env[62183]: DEBUG oslo_concurrency.lockutils [req-37bb3801-75de-4485-8b7b-82c44bfa4c50 req-a8bf109f-2de2-410e-93bd-22f2878bd4d5 service nova] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.112894] env[62183]: DEBUG oslo_concurrency.lockutils [req-37bb3801-75de-4485-8b7b-82c44bfa4c50 req-a8bf109f-2de2-410e-93bd-22f2878bd4d5 service nova] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.113296] env[62183]: DEBUG nova.compute.manager [req-37bb3801-75de-4485-8b7b-82c44bfa4c50 req-a8bf109f-2de2-410e-93bd-22f2878bd4d5 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] No waiting events found dispatching network-vif-plugged-d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1059.113513] env[62183]: WARNING nova.compute.manager [req-37bb3801-75de-4485-8b7b-82c44bfa4c50 req-a8bf109f-2de2-410e-93bd-22f2878bd4d5 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Received unexpected event network-vif-plugged-d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd for instance with vm_state building and task_state spawning. [ 1059.198943] env[62183]: DEBUG nova.network.neutron [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Successfully updated port: d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1059.423227] env[62183]: DEBUG oslo_vmware.api [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387495, 'name': ReconfigVM_Task, 'duration_secs': 0.222383} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.423523] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Reconfigured VM instance instance-0000005a to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1059.428123] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f44074a-1ccb-4594-b544-fa08f3c80a5d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.442188] env[62183]: DEBUG oslo_vmware.api [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1059.442188] env[62183]: value = "task-1387496" [ 1059.442188] env[62183]: _type = "Task" [ 1059.442188] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.449654] env[62183]: DEBUG oslo_vmware.api [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387496, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.459422] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.459543] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d4169be-8092-404d-8846-0ff1eb5457ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.464958] env[62183]: DEBUG oslo_vmware.api [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1059.464958] env[62183]: value = "task-1387497" [ 1059.464958] env[62183]: _type = "Task" [ 1059.464958] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.472930] env[62183]: DEBUG oslo_vmware.api [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387497, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.702385] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.702596] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.702713] env[62183]: DEBUG nova.network.neutron [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.952261] env[62183]: DEBUG oslo_vmware.api [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387496, 'name': ReconfigVM_Task, 'duration_secs': 0.139866} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.952544] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294514', 'volume_id': '786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'name': 'volume-786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8c4b68d6-fedd-408f-a449-aace7400014a', 'attached_at': '', 'detached_at': '', 'volume_id': '786b9d04-6ed3-4cb1-a773-6db867ff2adb', 'serial': '786b9d04-6ed3-4cb1-a773-6db867ff2adb'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1059.974463] env[62183]: DEBUG oslo_vmware.api [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387497, 'name': PowerOffVM_Task, 'duration_secs': 0.227221} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.974727] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.974907] env[62183]: DEBUG nova.compute.manager [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1059.975668] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2467e5ae-eaa4-46dd-b3b8-28911d77e0ed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.232636] env[62183]: DEBUG nova.network.neutron [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1060.359620] env[62183]: DEBUG nova.network.neutron [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance_info_cache with network_info: [{"id": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "address": "fa:16:3e:95:fa:9f", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2e95f9f-6c", "ovs_interfaceid": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.486890] env[62183]: DEBUG oslo_concurrency.lockutils [None req-1407354f-8e13-4578-a090-9d6a85241fc6 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "78db4224-11cb-468c-9c26-d470cac12b0e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.542s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.493395] env[62183]: DEBUG nova.objects.instance [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.862749] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.863265] env[62183]: DEBUG nova.compute.manager [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Instance network_info: |[{"id": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "address": "fa:16:3e:95:fa:9f", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2e95f9f-6c", "ovs_interfaceid": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1060.863889] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:fa:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8800a981-a89e-42e4-8be9-cace419ba9cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2e95f9f-6cee-45a9-80f2-3386ed4ca3bd', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1060.871655] env[62183]: DEBUG oslo.service.loopingcall [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.871896] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1060.872154] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-185c4272-f58c-4288-8b2f-a6faf5eda8c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.892179] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1060.892179] env[62183]: value = "task-1387498" [ 1060.892179] env[62183]: _type = "Task" [ 1060.892179] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.899827] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387498, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.146842] env[62183]: DEBUG nova.compute.manager [req-45e3b8bd-b672-4b54-9d3c-3bfe41ba3c95 req-a17b6c8f-e724-416e-a6a0-6d0e03985db7 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Received event network-changed-d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.147076] env[62183]: DEBUG nova.compute.manager [req-45e3b8bd-b672-4b54-9d3c-3bfe41ba3c95 req-a17b6c8f-e724-416e-a6a0-6d0e03985db7 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Refreshing instance network info cache due to event network-changed-d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1061.147334] env[62183]: DEBUG oslo_concurrency.lockutils [req-45e3b8bd-b672-4b54-9d3c-3bfe41ba3c95 req-a17b6c8f-e724-416e-a6a0-6d0e03985db7 service nova] Acquiring lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.147485] env[62183]: DEBUG oslo_concurrency.lockutils [req-45e3b8bd-b672-4b54-9d3c-3bfe41ba3c95 req-a17b6c8f-e724-416e-a6a0-6d0e03985db7 service nova] Acquired lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.147653] env[62183]: DEBUG nova.network.neutron [req-45e3b8bd-b672-4b54-9d3c-3bfe41ba3c95 req-a17b6c8f-e724-416e-a6a0-6d0e03985db7 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Refreshing network info cache for port d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1061.402312] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387498, 'name': CreateVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.469733] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "8c4b68d6-fedd-408f-a449-aace7400014a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.500040] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e08d3b5a-01af-41f5-94a5-e79284408855 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.203s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.501034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.031s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.501249] env[62183]: DEBUG nova.compute.manager [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1061.502314] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4b8df5-fd10-4664-bc64-e72e0dfd8112 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.509642] env[62183]: DEBUG nova.compute.manager [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62183) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1061.509642] env[62183]: DEBUG nova.objects.instance [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.547219] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "78db4224-11cb-468c-9c26-d470cac12b0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.547434] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "78db4224-11cb-468c-9c26-d470cac12b0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.547644] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "78db4224-11cb-468c-9c26-d470cac12b0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.547834] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "78db4224-11cb-468c-9c26-d470cac12b0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.548014] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "78db4224-11cb-468c-9c26-d470cac12b0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.550010] env[62183]: INFO nova.compute.manager [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Terminating instance [ 1061.551725] env[62183]: DEBUG nova.compute.manager [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1061.551919] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.552702] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad62dec1-a6f3-49f0-8e62-d927f305a82a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.559887] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1061.560118] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1ea5539-8d29-4c5c-83b3-802c46f8ff0d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.624356] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1061.624590] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1061.624782] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleting the datastore file [datastore2] 78db4224-11cb-468c-9c26-d470cac12b0e {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.625076] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-22bc8526-21e8-45e1-a5f5-457eb26f4651 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.630964] env[62183]: DEBUG oslo_vmware.api [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1061.630964] env[62183]: value = "task-1387500" [ 1061.630964] env[62183]: _type = "Task" [ 1061.630964] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.638451] env[62183]: DEBUG oslo_vmware.api [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387500, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.865220] env[62183]: DEBUG nova.network.neutron [req-45e3b8bd-b672-4b54-9d3c-3bfe41ba3c95 req-a17b6c8f-e724-416e-a6a0-6d0e03985db7 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updated VIF entry in instance network info cache for port d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1061.865583] env[62183]: DEBUG nova.network.neutron [req-45e3b8bd-b672-4b54-9d3c-3bfe41ba3c95 req-a17b6c8f-e724-416e-a6a0-6d0e03985db7 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance_info_cache with network_info: [{"id": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "address": "fa:16:3e:95:fa:9f", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2e95f9f-6c", "ovs_interfaceid": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.902989] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387498, 'name': CreateVM_Task, 'duration_secs': 0.519231} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.903157] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1061.903849] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.904038] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.904365] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1061.904616] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74425b59-8a37-425c-979a-f5481ef0d177 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.908664] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1061.908664] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52fcee76-0b39-55bd-adda-cb29627990fe" [ 1061.908664] env[62183]: _type = "Task" [ 1061.908664] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.916970] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52fcee76-0b39-55bd-adda-cb29627990fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.013940] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1062.014237] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-795f682c-abe8-45a3-95c4-4ca9a343cb00 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.022038] env[62183]: DEBUG oslo_vmware.api [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1062.022038] env[62183]: value = "task-1387501" [ 1062.022038] env[62183]: _type = "Task" [ 1062.022038] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.030347] env[62183]: DEBUG oslo_vmware.api [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387501, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.141392] env[62183]: DEBUG oslo_vmware.api [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387500, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151425} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.141392] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.141590] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1062.141852] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1062.142107] env[62183]: INFO nova.compute.manager [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1062.142412] env[62183]: DEBUG oslo.service.loopingcall [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.142620] env[62183]: DEBUG nova.compute.manager [-] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1062.142717] env[62183]: DEBUG nova.network.neutron [-] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1062.368322] env[62183]: DEBUG oslo_concurrency.lockutils [req-45e3b8bd-b672-4b54-9d3c-3bfe41ba3c95 req-a17b6c8f-e724-416e-a6a0-6d0e03985db7 service nova] Releasing lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.418416] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52fcee76-0b39-55bd-adda-cb29627990fe, 'name': SearchDatastore_Task, 'duration_secs': 0.00842} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.418510] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.418734] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1062.419132] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.419132] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.419310] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1062.419572] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1be8f8ad-0fdf-45b3-8e6c-9e6a7f6e346b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.427056] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1062.427263] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1062.427922] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75cfd7da-48a0-476b-b15e-513b7c3b7092 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.432330] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1062.432330] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52d68906-1564-b96b-9ab0-bdab19c39122" [ 1062.432330] env[62183]: _type = "Task" [ 1062.432330] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.439416] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d68906-1564-b96b-9ab0-bdab19c39122, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.531330] env[62183]: DEBUG oslo_vmware.api [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387501, 'name': PowerOffVM_Task, 'duration_secs': 0.18998} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.531609] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1062.531791] env[62183]: DEBUG nova.compute.manager [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1062.532857] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd287427-66d3-4dbb-b689-a6324a2cb428 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.894314] env[62183]: DEBUG nova.network.neutron [-] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.943391] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52d68906-1564-b96b-9ab0-bdab19c39122, 'name': SearchDatastore_Task, 'duration_secs': 0.00816} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.944245] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed326d09-2f55-48fa-81b1-6ff293f17fc8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.949544] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1062.949544] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52cf7da0-bb88-24d1-d503-9a6a6d73472a" [ 1062.949544] env[62183]: _type = "Task" [ 1062.949544] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.957228] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52cf7da0-bb88-24d1-d503-9a6a6d73472a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.044128] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0b38245c-cb52-4366-85be-eba6ae2d2f01 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.543s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.172407] env[62183]: DEBUG nova.compute.manager [req-64fa7558-883b-4269-86c0-8b664e04aca9 req-40180402-9f7d-4b28-84b5-c4db70f809d1 service nova] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Received event network-vif-deleted-6d72dc33-f503-4d8e-b8a9-1ea5979dbe78 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.397530] env[62183]: INFO nova.compute.manager [-] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Took 1.25 seconds to deallocate network for instance. [ 1063.461487] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52cf7da0-bb88-24d1-d503-9a6a6d73472a, 'name': SearchDatastore_Task, 'duration_secs': 0.00974} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.461766] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.462038] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 5aed3b2c-627c-4f33-8f22-950e758f822d/5aed3b2c-627c-4f33-8f22-950e758f822d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1063.462314] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49572e3c-1cc0-4d3a-956c-72b9282dac84 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.470152] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1063.470152] env[62183]: value = "task-1387502" [ 1063.470152] env[62183]: _type = "Task" [ 1063.470152] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.478425] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.847627] env[62183]: DEBUG nova.objects.instance [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.904112] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.904514] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.904789] env[62183]: DEBUG nova.objects.instance [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lazy-loading 'resources' on Instance uuid 78db4224-11cb-468c-9c26-d470cac12b0e {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.981063] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387502, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454807} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.981359] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 5aed3b2c-627c-4f33-8f22-950e758f822d/5aed3b2c-627c-4f33-8f22-950e758f822d.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1063.981649] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.981920] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f784582e-a03b-4900-9298-fbc488ceab7a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.989727] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1063.989727] env[62183]: value = "task-1387503" [ 1063.989727] env[62183]: _type = "Task" [ 1063.989727] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.000641] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387503, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.353251] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.353368] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.353523] env[62183]: DEBUG nova.network.neutron [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1064.353710] env[62183]: DEBUG nova.objects.instance [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'info_cache' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.487011] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d288533-c187-4e01-bdc3-3690a1b96a3e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.496695] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f334565c-8088-4c21-8f3b-e55d95ce3721 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.502283] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387503, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065404} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.502793] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1064.503508] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a13f59-844e-4693-9036-2d3ed7b82379 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.528855] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9112ab9a-d3df-4241-871c-280727640c08 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.548797] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 5aed3b2c-627c-4f33-8f22-950e758f822d/5aed3b2c-627c-4f33-8f22-950e758f822d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.549359] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89af8b35-104b-497c-87f7-225f0c685f45 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.566109] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa562d3d-ce30-4b2f-80bd-281e87994a24 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.570547] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1064.570547] env[62183]: value = "task-1387504" [ 1064.570547] env[62183]: _type = "Task" [ 1064.570547] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.584528] env[62183]: DEBUG nova.compute.provider_tree [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.589157] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387504, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.857499] env[62183]: DEBUG nova.objects.base [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Object Instance<8c4b68d6-fedd-408f-a449-aace7400014a> lazy-loaded attributes: flavor,info_cache {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1065.081662] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387504, 'name': ReconfigVM_Task, 'duration_secs': 0.284119} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.081956] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 5aed3b2c-627c-4f33-8f22-950e758f822d/5aed3b2c-627c-4f33-8f22-950e758f822d.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.082617] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b035f093-756b-4731-b94f-3983026b423b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.087129] env[62183]: DEBUG nova.scheduler.client.report [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1065.091690] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1065.091690] env[62183]: value = "task-1387505" [ 1065.091690] env[62183]: _type = "Task" [ 1065.091690] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.099629] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387505, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.562602] env[62183]: DEBUG nova.network.neutron [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updating instance_info_cache with network_info: [{"id": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "address": "fa:16:3e:10:9e:d2", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape301232c-2b", "ovs_interfaceid": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.591251] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.605367] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387505, 'name': Rename_Task, 'duration_secs': 0.141484} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.605814] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.606434] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3192850c-1cec-451e-a32e-6c58bcb930f9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.610039] env[62183]: INFO nova.scheduler.client.report [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted allocations for instance 78db4224-11cb-468c-9c26-d470cac12b0e [ 1065.616177] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1065.616177] env[62183]: value = "task-1387506" [ 1065.616177] env[62183]: _type = "Task" [ 1065.616177] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.627911] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387506, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.066113] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Releasing lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.122422] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c8912a70-c4b0-4f73-ad06-cc2788fd4432 tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "78db4224-11cb-468c-9c26-d470cac12b0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.575s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.128484] env[62183]: DEBUG oslo_vmware.api [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387506, 'name': PowerOnVM_Task, 'duration_secs': 0.428335} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.128766] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1066.128978] env[62183]: INFO nova.compute.manager [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Took 7.18 seconds to spawn the instance on the hypervisor. [ 1066.129169] env[62183]: DEBUG nova.compute.manager [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1066.129912] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1504a3a0-cd79-487c-bbca-0b51157d9b90 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.570398] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1066.570745] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1ff5882-db7a-4204-adb0-93e66b5df093 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.578138] env[62183]: DEBUG oslo_vmware.api [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1066.578138] env[62183]: value = "task-1387507" [ 1066.578138] env[62183]: _type = "Task" [ 1066.578138] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.586053] env[62183]: DEBUG oslo_vmware.api [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387507, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.646114] env[62183]: INFO nova.compute.manager [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Took 11.92 seconds to build instance. [ 1067.066220] env[62183]: DEBUG nova.compute.manager [req-24432ea9-991b-4373-9159-dfdbab573d39 req-f5fb32cd-d042-4f1d-aa1b-8cdf83b9ac20 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Received event network-changed-d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.066429] env[62183]: DEBUG nova.compute.manager [req-24432ea9-991b-4373-9159-dfdbab573d39 req-f5fb32cd-d042-4f1d-aa1b-8cdf83b9ac20 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Refreshing instance network info cache due to event network-changed-d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1067.066653] env[62183]: DEBUG oslo_concurrency.lockutils [req-24432ea9-991b-4373-9159-dfdbab573d39 req-f5fb32cd-d042-4f1d-aa1b-8cdf83b9ac20 service nova] Acquiring lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.066846] env[62183]: DEBUG oslo_concurrency.lockutils [req-24432ea9-991b-4373-9159-dfdbab573d39 req-f5fb32cd-d042-4f1d-aa1b-8cdf83b9ac20 service nova] Acquired lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.066968] env[62183]: DEBUG nova.network.neutron [req-24432ea9-991b-4373-9159-dfdbab573d39 req-f5fb32cd-d042-4f1d-aa1b-8cdf83b9ac20 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Refreshing network info cache for port d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.076890] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "88846abe-9e6e-4a88-bf6e-9978610213da" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.077173] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.089419] env[62183]: DEBUG oslo_vmware.api [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387507, 'name': PowerOnVM_Task, 'duration_secs': 0.394047} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.089684] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1067.089884] env[62183]: DEBUG nova.compute.manager [None req-ea5cb422-ec92-48c0-93b6-48c1460b7c1a tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.090653] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594e9076-1eb2-4a50-a8b5-796838b8823c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.148474] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c790290d-839b-4334-9e5c-341c67407c29 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.433s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.329307] env[62183]: DEBUG oslo_concurrency.lockutils [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.329649] env[62183]: DEBUG oslo_concurrency.lockutils [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.329911] env[62183]: DEBUG oslo_concurrency.lockutils [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.330144] env[62183]: DEBUG oslo_concurrency.lockutils [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.330340] env[62183]: DEBUG oslo_concurrency.lockutils [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.332734] env[62183]: INFO nova.compute.manager [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Terminating instance [ 1067.334629] env[62183]: DEBUG nova.compute.manager [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1067.334864] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.335803] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab26d5e-db37-4c14-86c7-3e6fb3e15862 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.343743] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.343971] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07e4d170-eb7c-425d-9be3-1143101d727e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.350154] env[62183]: DEBUG oslo_vmware.api [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1067.350154] env[62183]: value = "task-1387508" [ 1067.350154] env[62183]: _type = "Task" [ 1067.350154] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.357858] env[62183]: DEBUG oslo_vmware.api [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387508, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.580940] env[62183]: DEBUG nova.compute.utils [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1067.812412] env[62183]: DEBUG nova.network.neutron [req-24432ea9-991b-4373-9159-dfdbab573d39 req-f5fb32cd-d042-4f1d-aa1b-8cdf83b9ac20 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updated VIF entry in instance network info cache for port d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.813503] env[62183]: DEBUG nova.network.neutron [req-24432ea9-991b-4373-9159-dfdbab573d39 req-f5fb32cd-d042-4f1d-aa1b-8cdf83b9ac20 service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance_info_cache with network_info: [{"id": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "address": "fa:16:3e:95:fa:9f", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2e95f9f-6c", "ovs_interfaceid": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.859296] env[62183]: DEBUG oslo_vmware.api [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387508, 'name': PowerOffVM_Task, 'duration_secs': 0.215331} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.859570] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1067.859740] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1067.859996] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c43ebbc5-734c-4159-9c2c-8c22d8091785 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.920796] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1067.921063] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1067.921268] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleting the datastore file [datastore2] 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1067.921554] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fd23bfc-5019-4485-bb0f-9a5b065f7b51 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.930137] env[62183]: DEBUG oslo_vmware.api [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for the task: (returnval){ [ 1067.930137] env[62183]: value = "task-1387510" [ 1067.930137] env[62183]: _type = "Task" [ 1067.930137] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.937969] env[62183]: DEBUG oslo_vmware.api [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387510, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.084509] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.316041] env[62183]: DEBUG oslo_concurrency.lockutils [req-24432ea9-991b-4373-9159-dfdbab573d39 req-f5fb32cd-d042-4f1d-aa1b-8cdf83b9ac20 service nova] Releasing lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.440187] env[62183]: DEBUG oslo_vmware.api [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Task: {'id': task-1387510, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138368} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.440448] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.440643] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.440829] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.441019] env[62183]: INFO nova.compute.manager [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1068.441282] env[62183]: DEBUG oslo.service.loopingcall [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.441482] env[62183]: DEBUG nova.compute.manager [-] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1068.441578] env[62183]: DEBUG nova.network.neutron [-] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1069.091070] env[62183]: DEBUG nova.compute.manager [req-e7dd0a02-5ce2-47cf-873a-7b23fdf02708 req-2a1592de-a9f1-40a8-a0aa-0b7c84c651e9 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Received event network-vif-deleted-4e6d3d8d-6821-471a-bca5-29acccb74622 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.091401] env[62183]: INFO nova.compute.manager [req-e7dd0a02-5ce2-47cf-873a-7b23fdf02708 req-2a1592de-a9f1-40a8-a0aa-0b7c84c651e9 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Neutron deleted interface 4e6d3d8d-6821-471a-bca5-29acccb74622; detaching it from the instance and deleting it from the info cache [ 1069.091541] env[62183]: DEBUG nova.network.neutron [req-e7dd0a02-5ce2-47cf-873a-7b23fdf02708 req-2a1592de-a9f1-40a8-a0aa-0b7c84c651e9 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.148457] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.148457] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.148457] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.159082] env[62183]: DEBUG nova.network.neutron [-] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.160693] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "88846abe-9e6e-4a88-bf6e-9978610213da" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.160786] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.161020] env[62183]: INFO nova.compute.manager [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Attaching volume 6b106861-b013-4124-a712-16076f82c288 to /dev/sdb [ 1069.194637] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692270c2-7978-4c86-aa3c-8965d1023f43 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.201647] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93fe01c-86bb-4925-8ef6-69fa8b48c385 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.215480] env[62183]: DEBUG nova.virt.block_device [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating existing volume attachment record: fcf9a5df-47fd-4394-a93a-4a3125ab10ef {{(pid=62183) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1069.594220] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7975ef59-fcb9-4f69-9935-a82a4bce6826 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.603941] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff473c10-366d-4d0d-8308-56ec64e93b76 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.630117] env[62183]: DEBUG nova.compute.manager [req-e7dd0a02-5ce2-47cf-873a-7b23fdf02708 req-2a1592de-a9f1-40a8-a0aa-0b7c84c651e9 service nova] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Detach interface failed, port_id=4e6d3d8d-6821-471a-bca5-29acccb74622, reason: Instance 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1069.663320] env[62183]: INFO nova.compute.manager [-] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Took 1.22 seconds to deallocate network for instance. [ 1070.170555] env[62183]: DEBUG oslo_concurrency.lockutils [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.170889] env[62183]: DEBUG oslo_concurrency.lockutils [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.171140] env[62183]: DEBUG nova.objects.instance [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lazy-loading 'resources' on Instance uuid 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.741492] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef70391-0099-473d-a4f1-9593a0763636 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.748843] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5598ea-f460-433f-8b1e-e631917adadd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.777893] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec112087-ae15-4204-900f-415195aea4d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.784984] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ee111b-30b1-4283-adb1-584f81f98ced {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.797395] env[62183]: DEBUG nova.compute.provider_tree [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.300194] env[62183]: DEBUG nova.scheduler.client.report [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1071.649180] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.804886] env[62183]: DEBUG oslo_concurrency.lockutils [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.634s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.824232] env[62183]: INFO nova.scheduler.client.report [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Deleted allocations for instance 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e [ 1072.152404] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.152576] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.152733] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.152887] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1072.154187] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-598bc89f-14ac-46b2-a300-4186eee506b9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.162302] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32426bb-0467-46d8-a6d1-3666b6d23eb8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.176782] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7789dd-d0b3-4711-a2e4-e510b0f61cd5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.182918] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a5d533-76c9-4896-89fe-92483cdfa00b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.211017] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180385MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1072.211173] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.211369] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.331936] env[62183]: DEBUG oslo_concurrency.lockutils [None req-24da3523-2613-4d0b-bfce-2e2b44cb523b tempest-ServersTestJSON-1145816833 tempest-ServersTestJSON-1145816833-project-member] Lock "71fb3a2c-efd3-4f72-9997-136ee2ee4f6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.002s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.235965] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 8c4b68d6-fedd-408f-a449-aace7400014a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.236299] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 88846abe-9e6e-4a88-bf6e-9978610213da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.236503] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 1c180169-ebb5-4870-8e9d-925fe107b707 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.236605] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 5aed3b2c-627c-4f33-8f22-950e758f822d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1073.236893] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1073.237164] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1073.314288] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f613b5fb-98c4-4ba9-b175-30e2793f12d5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.323176] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871f562f-8821-4436-b5b1-66631325318c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.358483] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4056444e-24cc-473c-acea-db3c023bf0cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.366765] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be15341-bb1e-4eaa-8a10-efbbab7ff343 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.382902] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.762751] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Volume attach. Driver type: vmdk {{(pid=62183) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1073.763119] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294524', 'volume_id': '6b106861-b013-4124-a712-16076f82c288', 'name': 'volume-6b106861-b013-4124-a712-16076f82c288', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '88846abe-9e6e-4a88-bf6e-9978610213da', 'attached_at': '', 'detached_at': '', 'volume_id': '6b106861-b013-4124-a712-16076f82c288', 'serial': '6b106861-b013-4124-a712-16076f82c288'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1073.764076] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c6af2a-5b30-490c-95fd-4d5ed80ffadc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.781013] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b89d8c-ced6-4c8c-bc3c-eb257263ccf6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.806030] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] volume-6b106861-b013-4124-a712-16076f82c288/volume-6b106861-b013-4124-a712-16076f82c288.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.806515] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-579f2608-0993-45a5-b600-ff3ffc26d5a9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.825513] env[62183]: DEBUG oslo_vmware.api [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1073.825513] env[62183]: value = "task-1387515" [ 1073.825513] env[62183]: _type = "Task" [ 1073.825513] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.833977] env[62183]: DEBUG oslo_vmware.api [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387515, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.886293] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1074.337474] env[62183]: DEBUG oslo_vmware.api [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387515, 'name': ReconfigVM_Task, 'duration_secs': 0.464154} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.337850] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfigured VM instance instance-00000065 to attach disk [datastore2] volume-6b106861-b013-4124-a712-16076f82c288/volume-6b106861-b013-4124-a712-16076f82c288.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1074.343364] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-577aee31-a75e-4802-8a8b-c174ebb6709a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.359819] env[62183]: DEBUG oslo_vmware.api [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1074.359819] env[62183]: value = "task-1387516" [ 1074.359819] env[62183]: _type = "Task" [ 1074.359819] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.367487] env[62183]: DEBUG oslo_vmware.api [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387516, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.391627] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1074.391884] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.180s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.869581] env[62183]: DEBUG oslo_vmware.api [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387516, 'name': ReconfigVM_Task, 'duration_secs': 0.131997} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.869923] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294524', 'volume_id': '6b106861-b013-4124-a712-16076f82c288', 'name': 'volume-6b106861-b013-4124-a712-16076f82c288', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '88846abe-9e6e-4a88-bf6e-9978610213da', 'attached_at': '', 'detached_at': '', 'volume_id': '6b106861-b013-4124-a712-16076f82c288', 'serial': '6b106861-b013-4124-a712-16076f82c288'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1075.146826] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.147024] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.147144] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1075.147304] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Rebuilding the list of instances to heal {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1075.689069] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.689266] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquired lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.689385] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Forcefully refreshing network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1075.689553] env[62183]: DEBUG nova.objects.instance [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lazy-loading 'info_cache' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.906837] env[62183]: DEBUG nova.objects.instance [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'flavor' on Instance uuid 88846abe-9e6e-4a88-bf6e-9978610213da {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.412039] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fc9ce894-2d04-48e6-9f14-6a6dd2544337 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.251s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.760528] env[62183]: DEBUG oslo_concurrency.lockutils [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "1c180169-ebb5-4870-8e9d-925fe107b707" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.760802] env[62183]: DEBUG oslo_concurrency.lockutils [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.248633] env[62183]: DEBUG nova.compute.manager [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Stashing vm_state: active {{(pid=62183) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1077.263136] env[62183]: DEBUG nova.compute.utils [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1077.437973] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updating instance_info_cache with network_info: [{"id": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "address": "fa:16:3e:10:9e:d2", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape301232c-2b", "ovs_interfaceid": "e301232c-2b72-4bc5-b5f5-811b208bcc94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.765961] env[62183]: DEBUG oslo_concurrency.lockutils [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.894282] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.894571] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.940453] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Releasing lock "refresh_cache-8c4b68d6-fedd-408f-a449-aace7400014a" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.940660] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updated the network info_cache for instance {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1077.940858] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1077.941025] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1077.941187] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1077.941332] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1077.941464] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1077.941627] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1077.941750] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Cleaning up deleted instances with incomplete migration {{(pid=62183) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1078.398906] env[62183]: INFO nova.compute.claims [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.649783] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.831712] env[62183]: DEBUG oslo_concurrency.lockutils [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "1c180169-ebb5-4870-8e9d-925fe107b707" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.832078] env[62183]: DEBUG oslo_concurrency.lockutils [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.832247] env[62183]: INFO nova.compute.manager [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Attaching volume cf330958-5137-475c-8655-4a62b6985fcf to /dev/sdb [ 1078.861694] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c4049f-6756-4fa9-8557-82d7fc8d3f10 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.868666] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b150acdf-b3ad-48c5-8589-93447d3f60ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.881710] env[62183]: DEBUG nova.virt.block_device [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Updating existing volume attachment record: c4b1119e-7dbf-4c1b-865c-73a1c13009e7 {{(pid=62183) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1078.905316] env[62183]: INFO nova.compute.resource_tracker [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating resource usage from migration 37e2495e-91fa-4a54-8eaf-635a14877763 [ 1078.972878] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11d6f72-3c9e-490d-bba9-c54a2699fc77 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.980044] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc378728-d9ac-4c98-af2a-52dedcddcb53 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.008695] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd42eb9-0301-4798-9352-b71d189909b0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.018243] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e4cea4-af76-481b-95e8-eed6f5f53c81 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.029395] env[62183]: DEBUG nova.compute.provider_tree [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.155402] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.155589] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Cleaning up deleted instances {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1079.533105] env[62183]: DEBUG nova.scheduler.client.report [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1079.666729] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] There are 41 instances to clean {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1079.666993] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 78db4224-11cb-468c-9c26-d470cac12b0e] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.038669] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.144s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.039092] env[62183]: INFO nova.compute.manager [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Migrating [ 1080.169639] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 488d1c9b-fd4f-4b0d-bd05-9f42c2132b86] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.553524] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.553745] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.553925] env[62183]: DEBUG nova.network.neutron [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1080.673292] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: f51495aa-32b3-429c-8421-65f0a2587ea8] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.176327] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: ac58afb9-9f4c-4092-bcfc-b3ed1433a356] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.267317] env[62183]: DEBUG nova.network.neutron [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance_info_cache with network_info: [{"id": "be785bab-5120-4093-b24c-25e5c14608bf", "address": "fa:16:3e:da:b3:0d", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe785bab-51", "ovs_interfaceid": "be785bab-5120-4093-b24c-25e5c14608bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.680124] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: dbd34888-a333-40b4-ae9c-df5541b16704] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.771054] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.183433] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a5018b4e-016d-4c61-bfbf-2c039ab79499] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.686584] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: d8817d24-b0cb-4956-b195-cc417ae09fb4] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.190073] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 24f7c072-1209-4c6e-9a54-4e069f14f7d2] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.284843] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21e495a-7fea-4af6-8a45-dc81d6e9a42d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.305503] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance '88846abe-9e6e-4a88-bf6e-9978610213da' progress to 0 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.424811] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Volume attach. Driver type: vmdk {{(pid=62183) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1083.425075] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294525', 'volume_id': 'cf330958-5137-475c-8655-4a62b6985fcf', 'name': 'volume-cf330958-5137-475c-8655-4a62b6985fcf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1c180169-ebb5-4870-8e9d-925fe107b707', 'attached_at': '', 'detached_at': '', 'volume_id': 'cf330958-5137-475c-8655-4a62b6985fcf', 'serial': 'cf330958-5137-475c-8655-4a62b6985fcf'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1083.425963] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aecea96e-c27e-458d-aa8d-042d339754a8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.441915] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b419945c-5f6e-4a59-8373-29d95911cc6a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.465297] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] volume-cf330958-5137-475c-8655-4a62b6985fcf/volume-cf330958-5137-475c-8655-4a62b6985fcf.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.465546] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-930bd842-8465-40c7-83e6-1a4f228b44c5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.484916] env[62183]: DEBUG oslo_vmware.api [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1083.484916] env[62183]: value = "task-1387519" [ 1083.484916] env[62183]: _type = "Task" [ 1083.484916] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.494049] env[62183]: DEBUG oslo_vmware.api [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387519, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.693600] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 036037b3-9676-419c-ab87-e5a557438fbe] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.814246] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1083.814588] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eaa61e85-d33d-4820-b58f-6d1640b8c8fb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.822446] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1083.822446] env[62183]: value = "task-1387520" [ 1083.822446] env[62183]: _type = "Task" [ 1083.822446] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.829919] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.994264] env[62183]: DEBUG oslo_vmware.api [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387519, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.196832] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 143323db-da1c-4dd8-8f13-ed0af31e2027] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.332111] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387520, 'name': PowerOffVM_Task, 'duration_secs': 0.192888} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.332402] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1084.332588] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance '88846abe-9e6e-4a88-bf6e-9978610213da' progress to 17 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1084.494953] env[62183]: DEBUG oslo_vmware.api [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387519, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.700011] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: b6402e2f-1035-4c1b-9a1e-05b17d89e4bb] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.839041] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1084.839320] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1084.839426] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1084.839667] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1084.839838] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1084.839993] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1084.840219] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1084.840385] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1084.840554] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1084.840718] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1084.840889] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1084.845982] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-475b0018-b137-4b9b-8259-971935b89ec0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.861772] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1084.861772] env[62183]: value = "task-1387521" [ 1084.861772] env[62183]: _type = "Task" [ 1084.861772] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.869660] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387521, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.995581] env[62183]: DEBUG oslo_vmware.api [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387519, 'name': ReconfigVM_Task, 'duration_secs': 1.339482} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.995879] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Reconfigured VM instance instance-00000068 to attach disk [datastore2] volume-cf330958-5137-475c-8655-4a62b6985fcf/volume-cf330958-5137-475c-8655-4a62b6985fcf.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1085.000575] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd528763-f313-4ecf-973f-52886352a42d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.014306] env[62183]: DEBUG oslo_vmware.api [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1085.014306] env[62183]: value = "task-1387522" [ 1085.014306] env[62183]: _type = "Task" [ 1085.014306] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.022294] env[62183]: DEBUG oslo_vmware.api [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387522, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.203321] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 8ca36fc6-9c67-4b30-bf83-8cc8b1d084ec] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.371780] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387521, 'name': ReconfigVM_Task, 'duration_secs': 0.157746} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.372043] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance '88846abe-9e6e-4a88-bf6e-9978610213da' progress to 33 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1085.523440] env[62183]: DEBUG oslo_vmware.api [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387522, 'name': ReconfigVM_Task, 'duration_secs': 0.133789} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.523758] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294525', 'volume_id': 'cf330958-5137-475c-8655-4a62b6985fcf', 'name': 'volume-cf330958-5137-475c-8655-4a62b6985fcf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1c180169-ebb5-4870-8e9d-925fe107b707', 'attached_at': '', 'detached_at': '', 'volume_id': 'cf330958-5137-475c-8655-4a62b6985fcf', 'serial': 'cf330958-5137-475c-8655-4a62b6985fcf'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1085.706725] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 3a3ca703-2883-4aa9-a33f-326cc25d4838] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.879623] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1085.879936] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1085.880050] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1085.880248] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1085.880406] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1085.880556] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1085.880759] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1085.880918] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1085.881103] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1085.881274] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1085.881453] env[62183]: DEBUG nova.virt.hardware [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1085.886675] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1085.886958] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7c99cd6-f559-48d6-bf5f-d4a35f792c9f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.905192] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1085.905192] env[62183]: value = "task-1387523" [ 1085.905192] env[62183]: _type = "Task" [ 1085.905192] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.913040] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387523, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.210480] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: d30cfb69-1814-49b2-a29d-10cfbcc05722] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.418019] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387523, 'name': ReconfigVM_Task, 'duration_secs': 0.187269} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.418019] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1086.418019] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd8ebfd-042a-4b86-9b30-24c00d8e5d74 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.440444] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 88846abe-9e6e-4a88-bf6e-9978610213da/88846abe-9e6e-4a88-bf6e-9978610213da.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.440890] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2afed1ab-c4c2-407e-b84a-2ffaa9431720 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.458448] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1086.458448] env[62183]: value = "task-1387524" [ 1086.458448] env[62183]: _type = "Task" [ 1086.458448] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.465931] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387524, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.559783] env[62183]: DEBUG nova.objects.instance [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'flavor' on Instance uuid 1c180169-ebb5-4870-8e9d-925fe107b707 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.715747] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: ecdd6f2a-80bd-495d-b54a-cf09ace67a87] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.968844] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387524, 'name': ReconfigVM_Task, 'duration_secs': 0.308044} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.969144] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 88846abe-9e6e-4a88-bf6e-9978610213da/88846abe-9e6e-4a88-bf6e-9978610213da.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.969433] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance '88846abe-9e6e-4a88-bf6e-9978610213da' progress to 50 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1087.064668] env[62183]: DEBUG oslo_concurrency.lockutils [None req-863893bd-b4a5-434d-bc74-64ee56f6ec31 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.233s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.219378] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: c7f7547c-0fd3-417e-95ad-1bf3fc13bd2d] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.324400] env[62183]: DEBUG oslo_concurrency.lockutils [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "1c180169-ebb5-4870-8e9d-925fe107b707" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.324722] env[62183]: DEBUG oslo_concurrency.lockutils [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.475936] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf8c5cf-34d6-4e14-9354-17112df96230 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.496975] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1bbc4b-f05c-4e9a-9b32-bdc34aa2c9df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.517286] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance '88846abe-9e6e-4a88-bf6e-9978610213da' progress to 67 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1087.722736] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 7b8a64ad-afcb-4e7c-8c00-86f69d8d3631] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.827625] env[62183]: INFO nova.compute.manager [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Detaching volume cf330958-5137-475c-8655-4a62b6985fcf [ 1087.858494] env[62183]: INFO nova.virt.block_device [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Attempting to driver detach volume cf330958-5137-475c-8655-4a62b6985fcf from mountpoint /dev/sdb [ 1087.858739] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Volume detach. Driver type: vmdk {{(pid=62183) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1087.858934] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294525', 'volume_id': 'cf330958-5137-475c-8655-4a62b6985fcf', 'name': 'volume-cf330958-5137-475c-8655-4a62b6985fcf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1c180169-ebb5-4870-8e9d-925fe107b707', 'attached_at': '', 'detached_at': '', 'volume_id': 'cf330958-5137-475c-8655-4a62b6985fcf', 'serial': 'cf330958-5137-475c-8655-4a62b6985fcf'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1087.859842] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c82aa1-6113-46ca-bb75-1af8676dcd5c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.881176] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865b53ea-1f27-4317-b853-79ed6ed8ca19 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.888346] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047299db-369d-4cb2-a3a5-835927e1b76d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.907475] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e125420-afa8-4ad6-9def-4aa030267a43 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.923138] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] The volume has not been displaced from its original location: [datastore2] volume-cf330958-5137-475c-8655-4a62b6985fcf/volume-cf330958-5137-475c-8655-4a62b6985fcf.vmdk. No consolidation needed. {{(pid=62183) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1087.928305] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1087.928571] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6efb45d5-8863-436b-be7c-2041698ecb34 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.946477] env[62183]: DEBUG oslo_vmware.api [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1087.946477] env[62183]: value = "task-1387525" [ 1087.946477] env[62183]: _type = "Task" [ 1087.946477] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.954151] env[62183]: DEBUG oslo_vmware.api [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387525, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.226120] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: b89455d9-6d00-41ac-95ec-fb036b8c26c9] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.455911] env[62183]: DEBUG oslo_vmware.api [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387525, 'name': ReconfigVM_Task, 'duration_secs': 0.20249} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.456223] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1088.460961] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3af142e-9da0-4f75-9432-7f1617ffc32e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.475743] env[62183]: DEBUG oslo_vmware.api [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1088.475743] env[62183]: value = "task-1387526" [ 1088.475743] env[62183]: _type = "Task" [ 1088.475743] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.483422] env[62183]: DEBUG oslo_vmware.api [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387526, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.728512] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 28d279bd-206e-4ba0-bf84-2980ab41f38d] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.986055] env[62183]: DEBUG oslo_vmware.api [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387526, 'name': ReconfigVM_Task, 'duration_secs': 0.132616} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.986055] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294525', 'volume_id': 'cf330958-5137-475c-8655-4a62b6985fcf', 'name': 'volume-cf330958-5137-475c-8655-4a62b6985fcf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1c180169-ebb5-4870-8e9d-925fe107b707', 'attached_at': '', 'detached_at': '', 'volume_id': 'cf330958-5137-475c-8655-4a62b6985fcf', 'serial': 'cf330958-5137-475c-8655-4a62b6985fcf'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1089.155633] env[62183]: DEBUG nova.network.neutron [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Port be785bab-5120-4093-b24c-25e5c14608bf binding to destination host cpu-1 is already ACTIVE {{(pid=62183) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1089.232569] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 46aa1062-3fcd-4581-8647-32e1372e9491] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.530173] env[62183]: DEBUG nova.objects.instance [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'flavor' on Instance uuid 1c180169-ebb5-4870-8e9d-925fe107b707 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.736388] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 4de35fd3-7ef8-4a0b-b180-8b36e308e2e7] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.174751] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "88846abe-9e6e-4a88-bf6e-9978610213da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.174987] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.175179] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.239184] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 14ea1b33-ffbd-4c48-84ea-443c10d8317d] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.537833] env[62183]: DEBUG oslo_concurrency.lockutils [None req-48b5a23c-58a7-43dd-9fe8-764dfccc726d tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.213s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.742553] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 71fb3a2c-efd3-4f72-9997-136ee2ee4f6e] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.213274] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.213480] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.213669] env[62183]: DEBUG nova.network.neutron [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1091.245539] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: ba2cf670-6b0b-45c8-b0c3-8cdc7ab993de] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.553719] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "1c180169-ebb5-4870-8e9d-925fe107b707" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.553992] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.554235] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "1c180169-ebb5-4870-8e9d-925fe107b707-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.554451] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.554615] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.556863] env[62183]: INFO nova.compute.manager [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Terminating instance [ 1091.558713] env[62183]: DEBUG nova.compute.manager [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1091.558916] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1091.559764] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01c6947-1fcc-4d9e-88d9-349cef117e65 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.567493] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1091.567981] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b663a4d7-7832-4dac-8f77-00d5006c6492 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.573934] env[62183]: DEBUG oslo_vmware.api [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1091.573934] env[62183]: value = "task-1387527" [ 1091.573934] env[62183]: _type = "Task" [ 1091.573934] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.581061] env[62183]: DEBUG oslo_vmware.api [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.749014] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: be021940-5fff-429a-9b93-4ad238cdff77] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.955769] env[62183]: DEBUG nova.network.neutron [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance_info_cache with network_info: [{"id": "be785bab-5120-4093-b24c-25e5c14608bf", "address": "fa:16:3e:da:b3:0d", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe785bab-51", "ovs_interfaceid": "be785bab-5120-4093-b24c-25e5c14608bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.084311] env[62183]: DEBUG oslo_vmware.api [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387527, 'name': PowerOffVM_Task, 'duration_secs': 0.170849} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.085026] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1092.085026] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1092.085269] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3714e06-273b-4f6b-a814-8a908a377836 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.151075] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1092.151320] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1092.151511] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleting the datastore file [datastore1] 1c180169-ebb5-4870-8e9d-925fe107b707 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1092.151775] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f347aa19-c3ff-4f5d-b8f6-b07250966567 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.157892] env[62183]: DEBUG oslo_vmware.api [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1092.157892] env[62183]: value = "task-1387529" [ 1092.157892] env[62183]: _type = "Task" [ 1092.157892] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.164914] env[62183]: DEBUG oslo_vmware.api [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387529, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.252816] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 843d35ff-6871-4e2a-8e03-d7229a0d8246] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.459353] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.667068] env[62183]: DEBUG oslo_vmware.api [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387529, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128829} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.667353] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1092.667554] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1092.667739] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1092.667919] env[62183]: INFO nova.compute.manager [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1092.668187] env[62183]: DEBUG oslo.service.loopingcall [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.668407] env[62183]: DEBUG nova.compute.manager [-] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1092.668506] env[62183]: DEBUG nova.network.neutron [-] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1092.756260] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 1b0593f2-b712-4926-bfb8-30c92070ff6a] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.968459] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d735b83f-6355-45cc-a1c6-ec18eded4aa0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.975805] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d28a10e-5ce3-43e1-badf-ba07d289b840 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.150266] env[62183]: DEBUG nova.compute.manager [req-59f06b42-78d0-4183-b9f0-cec0a1caf0bd req-ba7f883e-27d0-47c6-b534-d373d562b85c service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Received event network-vif-deleted-78754a06-2366-4beb-bc13-95b404d4acf6 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.150266] env[62183]: INFO nova.compute.manager [req-59f06b42-78d0-4183-b9f0-cec0a1caf0bd req-ba7f883e-27d0-47c6-b534-d373d562b85c service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Neutron deleted interface 78754a06-2366-4beb-bc13-95b404d4acf6; detaching it from the instance and deleting it from the info cache [ 1093.150266] env[62183]: DEBUG nova.network.neutron [req-59f06b42-78d0-4183-b9f0-cec0a1caf0bd req-ba7f883e-27d0-47c6-b534-d373d562b85c service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.259682] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 7b5f50f8-5527-4f29-9fa6-6a598b77ef80] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.629121] env[62183]: DEBUG nova.network.neutron [-] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.651960] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a07f876-8734-4ed6-90a7-b6b763389d06 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.660813] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49cdd9a-64b3-4927-b297-3af95c5c3b3f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.686183] env[62183]: DEBUG nova.compute.manager [req-59f06b42-78d0-4183-b9f0-cec0a1caf0bd req-ba7f883e-27d0-47c6-b534-d373d562b85c service nova] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Detach interface failed, port_id=78754a06-2366-4beb-bc13-95b404d4acf6, reason: Instance 1c180169-ebb5-4870-8e9d-925fe107b707 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1093.762615] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: fb348784-62a0-4d1f-ac7f-f176f3da0dd9] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.074183] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5406cf-c695-439f-88eb-28d9b026245d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.095371] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba91a25d-617a-40e3-a6df-d23d1f34e006 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.101851] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance '88846abe-9e6e-4a88-bf6e-9978610213da' progress to 83 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1094.131993] env[62183]: INFO nova.compute.manager [-] [instance: 1c180169-ebb5-4870-8e9d-925fe107b707] Took 1.46 seconds to deallocate network for instance. [ 1094.265903] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 503786ca-dba4-43c1-9a25-9f1cbac9a6a4] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.608543] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1094.608828] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba52dbe1-dfbc-49a8-99c8-fae9279b7635 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.615301] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1094.615301] env[62183]: value = "task-1387530" [ 1094.615301] env[62183]: _type = "Task" [ 1094.615301] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.626616] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387530, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.637859] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.638225] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.638626] env[62183]: DEBUG nova.objects.instance [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'resources' on Instance uuid 1c180169-ebb5-4870-8e9d-925fe107b707 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.768632] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a948464b-63aa-4bc8-9885-228049e96d37] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.126044] env[62183]: DEBUG oslo_vmware.api [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387530, 'name': PowerOnVM_Task, 'duration_secs': 0.393436} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.126357] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1095.126548] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ffcf94a1-3ba7-40d0-88bc-50c28e8db293 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance '88846abe-9e6e-4a88-bf6e-9978610213da' progress to 100 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1095.210387] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32a7a10-40b8-4b3c-90b1-cd6e916f9355 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.218182] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8155b4-f378-4660-9a71-e5f06f9e5e04 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.250541] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283334e3-3288-4249-9ccd-3b59016e4cf0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.257531] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb1809c-8ac1-4918-82e3-743721ef543e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.270683] env[62183]: DEBUG nova.compute.provider_tree [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1095.272137] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: e3145bd7-85b2-4cc7-9d97-3e36a59b89cb] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.777811] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 4fa0beb2-3f36-48f4-996c-1da0bb891f93] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.793812] env[62183]: ERROR nova.scheduler.client.report [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [req-3ce78fd7-a64b-49fb-b080-3e5a64c077b0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3ce78fd7-a64b-49fb-b080-3e5a64c077b0"}]} [ 1095.808595] env[62183]: DEBUG nova.scheduler.client.report [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1095.820604] env[62183]: DEBUG nova.scheduler.client.report [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1095.820817] env[62183]: DEBUG nova.compute.provider_tree [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1095.830478] env[62183]: DEBUG nova.scheduler.client.report [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1095.846663] env[62183]: DEBUG nova.scheduler.client.report [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1095.906239] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587be259-8de2-42b0-8224-9a4ee3f6a787 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.913961] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615e7848-ed4c-4460-b458-7da83bb898e9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.942955] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b81f1d7-934a-465f-aed8-a158f5c3356b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.950491] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90af2074-0611-42aa-bac0-d28c3560f133 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.965433] env[62183]: DEBUG nova.compute.provider_tree [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1096.280746] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: a5dc320c-67b5-4b8e-b9f6-16ec3bf93e5a] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.494550] env[62183]: DEBUG nova.scheduler.client.report [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 136 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1096.494833] env[62183]: DEBUG nova.compute.provider_tree [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 136 to 137 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1096.495031] env[62183]: DEBUG nova.compute.provider_tree [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1096.783351] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 850d40d8-565a-49a2-a27f-3de2a8dc7e30] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.981452] env[62183]: DEBUG nova.network.neutron [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Port be785bab-5120-4093-b24c-25e5c14608bf binding to destination host cpu-1 is already ACTIVE {{(pid=62183) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1096.981753] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.981912] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.982098] env[62183]: DEBUG nova.network.neutron [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1096.999802] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.362s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.020298] env[62183]: INFO nova.scheduler.client.report [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleted allocations for instance 1c180169-ebb5-4870-8e9d-925fe107b707 [ 1097.286386] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 39683931-d0ff-4a5c-a4a2-792230ab0e3d] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.529339] env[62183]: DEBUG oslo_concurrency.lockutils [None req-4f526c5f-b1c8-43b2-b562-09fa7d3168f7 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "1c180169-ebb5-4870-8e9d-925fe107b707" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.975s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.726642] env[62183]: DEBUG nova.network.neutron [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance_info_cache with network_info: [{"id": "be785bab-5120-4093-b24c-25e5c14608bf", "address": "fa:16:3e:da:b3:0d", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe785bab-51", "ovs_interfaceid": "be785bab-5120-4093-b24c-25e5c14608bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.789731] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 68e791b4-61db-4b6b-a30a-ccf5d6657643] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.229773] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.292665] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: ac555ffc-ce4e-4650-97fd-c26a3246fe4b] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.733763] env[62183]: DEBUG nova.compute.manager [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62183) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1098.795606] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 7465df88-c68c-49d3-9a91-6fff0d06957b] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.298920] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 07899178-b53c-4b0e-877d-11774286212f] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.611213] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.611485] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.802763] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 4af88a37-5aa2-47af-9dd9-8233b1bbf077] Instance has had 0 of 5 cleanup attempts {{(pid=62183) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.831197] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.831492] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.114237] env[62183]: DEBUG nova.compute.manager [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1100.333974] env[62183]: DEBUG nova.objects.instance [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'migration_context' on Instance uuid 88846abe-9e6e-4a88-bf6e-9978610213da {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.635852] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.902985] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4293b06a-8d3b-4360-bfed-95f6a51edd79 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.911266] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c213409-7994-46da-862b-bbfaecb8fe36 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.942651] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ae0a50-ea91-49ab-b98b-d53adbf3abd5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.949844] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb14535-78b8-4433-8160-5f7966b1b9af {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.962412] env[62183]: DEBUG nova.compute.provider_tree [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.465444] env[62183]: DEBUG nova.scheduler.client.report [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1102.479705] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.648s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.500317] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.864s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.502130] env[62183]: INFO nova.compute.claims [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1103.578403] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac60460-d81d-4c11-a0d5-9755b3877970 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.585885] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c813eb3-ba09-4d4f-b0cb-0948b1078d8d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.614680] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4983575c-2f31-4914-988a-b10680e5252c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.621781] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6093e1df-41e8-4188-9b1d-2c6c5c9f4a48 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.634623] env[62183]: DEBUG nova.compute.provider_tree [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.034190] env[62183]: INFO nova.compute.manager [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Swapping old allocation on dict_keys(['09c07e5d-2ed9-41c2-be62-db0f731d0b87']) held by migration 37e2495e-91fa-4a54-8eaf-635a14877763 for instance [ 1104.055596] env[62183]: DEBUG nova.scheduler.client.report [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Overwriting current allocation {'allocations': {'09c07e5d-2ed9-41c2-be62-db0f731d0b87': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 138}}, 'project_id': '88022032e6e04a4f96bc49c2ca5ede29', 'user_id': 'f4b92513f568466e81075af3fa4604fa', 'consumer_generation': 1} on consumer 88846abe-9e6e-4a88-bf6e-9978610213da {{(pid=62183) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2033}} [ 1104.133467] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.133685] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.133838] env[62183]: DEBUG nova.network.neutron [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1104.137274] env[62183]: DEBUG nova.scheduler.client.report [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1104.362615] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "8c4b68d6-fedd-408f-a449-aace7400014a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.362947] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.363151] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "8c4b68d6-fedd-408f-a449-aace7400014a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.363343] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.363518] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.366044] env[62183]: INFO nova.compute.manager [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Terminating instance [ 1104.367682] env[62183]: DEBUG nova.compute.manager [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1104.367890] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1104.368804] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdcf354c-fbb0-41da-a0d2-9d61c769e403 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.376454] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1104.376685] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2bfd6b38-5332-4146-9607-28357717d306 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.385121] env[62183]: DEBUG oslo_vmware.api [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1104.385121] env[62183]: value = "task-1387531" [ 1104.385121] env[62183]: _type = "Task" [ 1104.385121] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.393353] env[62183]: DEBUG oslo_vmware.api [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387531, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.641738] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.141s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.642282] env[62183]: DEBUG nova.compute.manager [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1104.854500] env[62183]: DEBUG nova.network.neutron [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance_info_cache with network_info: [{"id": "be785bab-5120-4093-b24c-25e5c14608bf", "address": "fa:16:3e:da:b3:0d", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe785bab-51", "ovs_interfaceid": "be785bab-5120-4093-b24c-25e5c14608bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.894783] env[62183]: DEBUG oslo_vmware.api [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387531, 'name': PowerOffVM_Task, 'duration_secs': 0.167246} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.895026] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1104.895217] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1104.895496] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-591a61f5-3c12-4efa-af65-4bf0d07daaa1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.957204] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1104.957581] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1104.957725] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Deleting the datastore file [datastore1] 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1104.957987] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9948b23-efcd-4074-b224-b1aa724d2b86 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.964437] env[62183]: DEBUG oslo_vmware.api [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1104.964437] env[62183]: value = "task-1387533" [ 1104.964437] env[62183]: _type = "Task" [ 1104.964437] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.973027] env[62183]: DEBUG oslo_vmware.api [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387533, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.148179] env[62183]: DEBUG nova.compute.utils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1105.149546] env[62183]: DEBUG nova.compute.manager [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1105.149725] env[62183]: DEBUG nova.network.neutron [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1105.187982] env[62183]: DEBUG nova.policy [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ae1837d5db145278417f7cdd55a3fea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06e81bd1a81d4009ae2a75fe819f9b7c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1105.321009] env[62183]: DEBUG nova.compute.manager [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Stashing vm_state: active {{(pid=62183) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1105.358179] env[62183]: DEBUG oslo_concurrency.lockutils [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-88846abe-9e6e-4a88-bf6e-9978610213da" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.359301] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7814b6b-03ec-4eb9-a214-65d0420d6008 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.367058] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52cc85a4-483c-4b95-bb85-6ed5a67244ac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.473813] env[62183]: DEBUG oslo_vmware.api [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387533, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133868} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.474507] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1105.474507] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1105.474507] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1105.474617] env[62183]: INFO nova.compute.manager [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1105.474854] env[62183]: DEBUG oslo.service.loopingcall [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1105.475060] env[62183]: DEBUG nova.compute.manager [-] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1105.475158] env[62183]: DEBUG nova.network.neutron [-] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1105.541122] env[62183]: DEBUG nova.network.neutron [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Successfully created port: c7b92447-a61f-4d71-b39f-28af011a2142 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1105.652581] env[62183]: DEBUG nova.compute.manager [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1105.841739] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.842032] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.277177] env[62183]: DEBUG nova.compute.manager [req-dce8820c-0ade-4061-a501-661204d2bd77 req-362fa873-9ade-4f27-a286-cd9203bf7857 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Received event network-vif-deleted-e301232c-2b72-4bc5-b5f5-811b208bcc94 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.277443] env[62183]: INFO nova.compute.manager [req-dce8820c-0ade-4061-a501-661204d2bd77 req-362fa873-9ade-4f27-a286-cd9203bf7857 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Neutron deleted interface e301232c-2b72-4bc5-b5f5-811b208bcc94; detaching it from the instance and deleting it from the info cache [ 1106.277631] env[62183]: DEBUG nova.network.neutron [req-dce8820c-0ade-4061-a501-661204d2bd77 req-362fa873-9ade-4f27-a286-cd9203bf7857 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.347713] env[62183]: INFO nova.compute.claims [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1106.450724] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1106.451050] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-629b3096-0958-44b4-b808-018bf8716a71 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.460023] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1106.460023] env[62183]: value = "task-1387534" [ 1106.460023] env[62183]: _type = "Task" [ 1106.460023] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.467812] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387534, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.662305] env[62183]: DEBUG nova.compute.manager [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1106.686997] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1106.687252] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1106.687496] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1106.687703] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1106.687855] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1106.688014] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1106.688235] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1106.688404] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1106.688623] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1106.688801] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1106.688981] env[62183]: DEBUG nova.virt.hardware [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.689853] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e93b12-de8e-4308-b0f1-3ac340d35ed2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.698205] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1fe918-76f5-40ae-b62f-e1ad21f79388 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.757961] env[62183]: DEBUG nova.network.neutron [-] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.781303] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e51e30f6-c5d6-4115-a10b-e28392b11d0c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.790683] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9834769-1f5e-48a9-a147-fe54b7f0c9ed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.816993] env[62183]: DEBUG nova.compute.manager [req-dce8820c-0ade-4061-a501-661204d2bd77 req-362fa873-9ade-4f27-a286-cd9203bf7857 service nova] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Detach interface failed, port_id=e301232c-2b72-4bc5-b5f5-811b208bcc94, reason: Instance 8c4b68d6-fedd-408f-a449-aace7400014a could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1106.854314] env[62183]: INFO nova.compute.resource_tracker [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating resource usage from migration 27011ab2-e2da-49dc-967a-0a676032d5ed [ 1106.926646] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c33b8f4-fa76-498a-98ff-4e28e9650aaa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.934396] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1768ece-2526-4f55-8b10-9a41932a470f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.971316] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f32f4b5-416e-4cf1-9b66-1bd3fe13fa8b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.980746] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387534, 'name': PowerOffVM_Task, 'duration_secs': 0.199914} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.983471] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1106.984351] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1106.984950] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1106.984950] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1106.985159] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1106.985379] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1106.985601] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1106.985895] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1106.986140] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1106.986385] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1106.986620] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1106.986869] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.994723] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e034f3c-bfaf-4ef9-9386-487e33dfccc1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.012096] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e15f95fb-3792-4b93-84d7-c0ca9fbddc94 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.029173] env[62183]: DEBUG nova.compute.provider_tree [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1107.034044] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1107.034044] env[62183]: value = "task-1387535" [ 1107.034044] env[62183]: _type = "Task" [ 1107.034044] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.039117] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387535, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.262528] env[62183]: INFO nova.compute.manager [-] [instance: 8c4b68d6-fedd-408f-a449-aace7400014a] Took 1.79 seconds to deallocate network for instance. [ 1107.391054] env[62183]: DEBUG nova.network.neutron [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Successfully updated port: c7b92447-a61f-4d71-b39f-28af011a2142 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1107.534457] env[62183]: DEBUG nova.scheduler.client.report [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1107.549967] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387535, 'name': ReconfigVM_Task, 'duration_secs': 0.144276} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.551405] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e727a037-5eb5-4d79-a3ec-cd31e97c0ea1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.574859] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1107.575160] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1107.575645] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1107.575727] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1107.575969] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1107.576250] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1107.576670] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1107.577282] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1107.577282] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1107.577563] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1107.577865] env[62183]: DEBUG nova.virt.hardware [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1107.579440] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e26b95d7-2431-41ae-8901-51bf8c8d3089 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.585178] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1107.585178] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52629f52-c88d-1517-07a9-3da41edf3c3b" [ 1107.585178] env[62183]: _type = "Task" [ 1107.585178] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.592956] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52629f52-c88d-1517-07a9-3da41edf3c3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.768688] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.894078] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "refresh_cache-632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.894255] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "refresh_cache-632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.894385] env[62183]: DEBUG nova.network.neutron [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1108.044758] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.203s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.044999] env[62183]: INFO nova.compute.manager [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Migrating [ 1108.054671] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.286s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.054912] env[62183]: DEBUG nova.objects.instance [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'resources' on Instance uuid 8c4b68d6-fedd-408f-a449-aace7400014a {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.094783] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52629f52-c88d-1517-07a9-3da41edf3c3b, 'name': SearchDatastore_Task, 'duration_secs': 0.008283} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.100122] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1108.100395] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2165453-2d77-4e72-9f71-4628baaaeed1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.118338] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1108.118338] env[62183]: value = "task-1387536" [ 1108.118338] env[62183]: _type = "Task" [ 1108.118338] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.125725] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387536, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.306303] env[62183]: DEBUG nova.compute.manager [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Received event network-vif-plugged-c7b92447-a61f-4d71-b39f-28af011a2142 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.306637] env[62183]: DEBUG oslo_concurrency.lockutils [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] Acquiring lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.306979] env[62183]: DEBUG oslo_concurrency.lockutils [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.307269] env[62183]: DEBUG oslo_concurrency.lockutils [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.307563] env[62183]: DEBUG nova.compute.manager [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] No waiting events found dispatching network-vif-plugged-c7b92447-a61f-4d71-b39f-28af011a2142 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1108.307836] env[62183]: WARNING nova.compute.manager [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Received unexpected event network-vif-plugged-c7b92447-a61f-4d71-b39f-28af011a2142 for instance with vm_state building and task_state spawning. [ 1108.308130] env[62183]: DEBUG nova.compute.manager [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Received event network-changed-c7b92447-a61f-4d71-b39f-28af011a2142 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.308404] env[62183]: DEBUG nova.compute.manager [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Refreshing instance network info cache due to event network-changed-c7b92447-a61f-4d71-b39f-28af011a2142. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1108.308667] env[62183]: DEBUG oslo_concurrency.lockutils [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] Acquiring lock "refresh_cache-632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.426815] env[62183]: DEBUG nova.network.neutron [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1108.555747] env[62183]: DEBUG nova.network.neutron [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Updating instance_info_cache with network_info: [{"id": "c7b92447-a61f-4d71-b39f-28af011a2142", "address": "fa:16:3e:f9:e5:a6", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7b92447-a6", "ovs_interfaceid": "c7b92447-a61f-4d71-b39f-28af011a2142", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.560714] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.560883] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.561069] env[62183]: DEBUG nova.network.neutron [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1108.624376] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d6fdb5-a649-4c1c-a7ae-73669f93f8b6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.629658] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387536, 'name': ReconfigVM_Task, 'duration_secs': 0.195213} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.630246] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1108.630996] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edfa953e-aa20-4462-aae5-44a533959a13 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.635894] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de06e41c-f367-480a-96fc-ab4fd4048832 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.659660] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 88846abe-9e6e-4a88-bf6e-9978610213da/88846abe-9e6e-4a88-bf6e-9978610213da.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1108.660367] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3d59267-5381-487d-a3f3-6efba2e377da {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.696499] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18723abb-b495-4e60-993d-1bf42616f3ea {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.704721] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fab6b5-6f6f-4859-aa5f-c1b9ad0aba70 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.708599] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1108.708599] env[62183]: value = "task-1387537" [ 1108.708599] env[62183]: _type = "Task" [ 1108.708599] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.719989] env[62183]: DEBUG nova.compute.provider_tree [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.725700] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.060529] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "refresh_cache-632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.060933] env[62183]: DEBUG nova.compute.manager [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Instance network_info: |[{"id": "c7b92447-a61f-4d71-b39f-28af011a2142", "address": "fa:16:3e:f9:e5:a6", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7b92447-a6", "ovs_interfaceid": "c7b92447-a61f-4d71-b39f-28af011a2142", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1109.061288] env[62183]: DEBUG oslo_concurrency.lockutils [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] Acquired lock "refresh_cache-632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.061478] env[62183]: DEBUG nova.network.neutron [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Refreshing network info cache for port c7b92447-a61f-4d71-b39f-28af011a2142 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1109.063023] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:e5:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '205fb402-8eaf-4b61-8f57-8f216024179a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7b92447-a61f-4d71-b39f-28af011a2142', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1109.070714] env[62183]: DEBUG oslo.service.loopingcall [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.073529] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1109.073969] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ff5b2bc-1b92-40b9-b8f6-233fc0fc5727 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.093280] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1109.093280] env[62183]: value = "task-1387538" [ 1109.093280] env[62183]: _type = "Task" [ 1109.093280] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.100717] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387538, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.219539] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387537, 'name': ReconfigVM_Task, 'duration_secs': 0.278031} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.219837] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 88846abe-9e6e-4a88-bf6e-9978610213da/88846abe-9e6e-4a88-bf6e-9978610213da.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1109.220640] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8863d579-4d9a-41db-814e-9d65469209c6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.223433] env[62183]: DEBUG nova.scheduler.client.report [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1109.246790] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aeeaa67-1f1d-419b-ac71-2df47753f13e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.269689] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03afa58-abf1-4310-85f0-02eb10758d03 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.290764] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8c25be-0b4d-4202-b7fb-411e9466d498 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.297965] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1109.298255] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87068d2d-dbe4-4a3f-9dd2-6034218dc951 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.303861] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1109.303861] env[62183]: value = "task-1387539" [ 1109.303861] env[62183]: _type = "Task" [ 1109.303861] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.304928] env[62183]: DEBUG nova.network.neutron [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance_info_cache with network_info: [{"id": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "address": "fa:16:3e:95:fa:9f", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2e95f9f-6c", "ovs_interfaceid": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.314824] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387539, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.606956] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387538, 'name': CreateVM_Task, 'duration_secs': 0.368794} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.610292] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1109.611218] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.611533] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.611985] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1109.612740] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-439a7a92-2c46-4a4c-a65e-1c7123849e75 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.618407] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1109.618407] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52546ece-8768-b12d-ca02-346ee17b48b2" [ 1109.618407] env[62183]: _type = "Task" [ 1109.618407] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.631797] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52546ece-8768-b12d-ca02-346ee17b48b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.730614] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.676s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.750634] env[62183]: INFO nova.scheduler.client.report [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Deleted allocations for instance 8c4b68d6-fedd-408f-a449-aace7400014a [ 1109.780275] env[62183]: DEBUG nova.network.neutron [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Updated VIF entry in instance network info cache for port c7b92447-a61f-4d71-b39f-28af011a2142. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1109.780842] env[62183]: DEBUG nova.network.neutron [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Updating instance_info_cache with network_info: [{"id": "c7b92447-a61f-4d71-b39f-28af011a2142", "address": "fa:16:3e:f9:e5:a6", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7b92447-a6", "ovs_interfaceid": "c7b92447-a61f-4d71-b39f-28af011a2142", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.813694] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.817913] env[62183]: DEBUG oslo_vmware.api [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387539, 'name': PowerOnVM_Task, 'duration_secs': 0.48184} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.817913] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1110.129432] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52546ece-8768-b12d-ca02-346ee17b48b2, 'name': SearchDatastore_Task, 'duration_secs': 0.00922} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.129779] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.129984] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1110.130242] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.130398] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.130582] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1110.130850] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97f98a15-3679-483a-aef3-907e82175dcb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.138841] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1110.139123] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1110.139802] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5352283e-3b3e-4323-9dc2-91396c219667 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.144472] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1110.144472] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]529a06f0-aa44-4a06-8120-a6edd7c946c7" [ 1110.144472] env[62183]: _type = "Task" [ 1110.144472] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.151631] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529a06f0-aa44-4a06-8120-a6edd7c946c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.260303] env[62183]: DEBUG oslo_concurrency.lockutils [None req-d16cc520-f6b8-41e2-9942-af6188b6ba99 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "8c4b68d6-fedd-408f-a449-aace7400014a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.897s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.283616] env[62183]: DEBUG oslo_concurrency.lockutils [req-1a055013-91e7-4a9e-bf16-f5f5632bff68 req-640b22ba-313b-4339-8fae-c7de1b36ca7a service nova] Releasing lock "refresh_cache-632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.655116] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529a06f0-aa44-4a06-8120-a6edd7c946c7, 'name': SearchDatastore_Task, 'duration_secs': 0.00791} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.655970] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-281a9836-dbda-4949-a2b8-b01626c96080 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.660724] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1110.660724] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52c30f96-3b66-6895-2b7b-8b45321b1e54" [ 1110.660724] env[62183]: _type = "Task" [ 1110.660724] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.667880] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c30f96-3b66-6895-2b7b-8b45321b1e54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.960135] env[62183]: INFO nova.compute.manager [None req-f2019f6f-dd52-438e-ae0a-1a70b734c3eb tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance to original state: 'active' [ 1111.171208] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52c30f96-3b66-6895-2b7b-8b45321b1e54, 'name': SearchDatastore_Task, 'duration_secs': 0.009956} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.171499] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.171749] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11/632ccdc6-d5fa-4e42-8273-c2e46eb1bf11.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1111.172031] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07439b48-3c19-4259-82b6-e95a250e0ce1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.178478] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1111.178478] env[62183]: value = "task-1387540" [ 1111.178478] env[62183]: _type = "Task" [ 1111.178478] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.185675] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.326953] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65510441-ae72-462b-a685-6c558568a30e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.356793] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance '5aed3b2c-627c-4f33-8f22-950e758f822d' progress to 0 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1111.688113] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471057} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.688408] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11/632ccdc6-d5fa-4e42-8273-c2e46eb1bf11.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1111.688649] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1111.688947] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08a7711a-3e07-48b4-92b5-d984a859616f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.695251] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1111.695251] env[62183]: value = "task-1387541" [ 1111.695251] env[62183]: _type = "Task" [ 1111.695251] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.703252] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387541, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.864838] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.865193] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00e91cee-270c-4730-99b8-e6b66548a847 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.872517] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1111.872517] env[62183]: value = "task-1387542" [ 1111.872517] env[62183]: _type = "Task" [ 1111.872517] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.880578] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.205265] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387541, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066237} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.205573] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1112.206353] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f006e2a9-d3ef-4b20-8902-19052bafc4d6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.230055] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11/632ccdc6-d5fa-4e42-8273-c2e46eb1bf11.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.230055] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-560cec87-a8e1-4d16-a8c6-97b893dded9b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.249961] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1112.249961] env[62183]: value = "task-1387543" [ 1112.249961] env[62183]: _type = "Task" [ 1112.249961] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.258215] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387543, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.382360] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.382579] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.387057] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387542, 'name': PowerOffVM_Task, 'duration_secs': 0.184969} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.387553] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1112.387763] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance '5aed3b2c-627c-4f33-8f22-950e758f822d' progress to 17 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1112.760505] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387543, 'name': ReconfigVM_Task, 'duration_secs': 0.262944} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.760856] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11/632ccdc6-d5fa-4e42-8273-c2e46eb1bf11.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.761513] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7a790d4-f7a2-4f59-b9b8-107deeac2789 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.767604] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1112.767604] env[62183]: value = "task-1387544" [ 1112.767604] env[62183]: _type = "Task" [ 1112.767604] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.775671] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387544, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.888531] env[62183]: DEBUG nova.compute.manager [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1112.893096] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:04Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1112.893364] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1112.893551] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1112.893763] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1112.893917] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1112.894081] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1112.894291] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1112.894453] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1112.894619] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1112.894783] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1112.894956] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1112.900407] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-376d2b4c-fff9-40f6-b208-c1f22269a6ee {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.916104] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1112.916104] env[62183]: value = "task-1387545" [ 1112.916104] env[62183]: _type = "Task" [ 1112.916104] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.924397] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387545, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.935926] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "88846abe-9e6e-4a88-bf6e-9978610213da" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.936198] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.936440] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "88846abe-9e6e-4a88-bf6e-9978610213da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.936716] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.936923] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.939285] env[62183]: INFO nova.compute.manager [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Terminating instance [ 1112.941584] env[62183]: DEBUG nova.compute.manager [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1112.941822] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1112.942146] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0679ce56-1968-485c-ab91-eddd57db4910 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.949099] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1112.949099] env[62183]: value = "task-1387546" [ 1112.949099] env[62183]: _type = "Task" [ 1112.949099] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.957247] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387546, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.286131] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387544, 'name': Rename_Task, 'duration_secs': 0.133802} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.286634] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1113.286752] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-974d4047-aa00-45ae-b961-3180560b2774 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.293847] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1113.293847] env[62183]: value = "task-1387547" [ 1113.293847] env[62183]: _type = "Task" [ 1113.293847] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.309917] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387547, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.421250] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.421516] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.423154] env[62183]: INFO nova.compute.claims [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1113.429358] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387545, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.457997] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387546, 'name': PowerOffVM_Task, 'duration_secs': 0.444611} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.458317] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1113.458557] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Volume detach. Driver type: vmdk {{(pid=62183) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1113.458767] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294524', 'volume_id': '6b106861-b013-4124-a712-16076f82c288', 'name': 'volume-6b106861-b013-4124-a712-16076f82c288', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '88846abe-9e6e-4a88-bf6e-9978610213da', 'attached_at': '2024-10-31T11:15:32.000000', 'detached_at': '', 'volume_id': '6b106861-b013-4124-a712-16076f82c288', 'serial': '6b106861-b013-4124-a712-16076f82c288'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1113.459617] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5ad681-c9e4-497a-94fc-d82c554a2f52 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.482404] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6919bf75-4fab-4a26-9375-07af291bd173 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.489070] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57490d3b-d0b7-4ed4-a30d-c35f1354cc5a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.511009] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6385a3fc-76e0-4642-9b6b-b2d37ff03d7c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.525188] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] The volume has not been displaced from its original location: [datastore2] volume-6b106861-b013-4124-a712-16076f82c288/volume-6b106861-b013-4124-a712-16076f82c288.vmdk. No consolidation needed. {{(pid=62183) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1113.530344] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfiguring VM instance instance-00000065 to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1113.530927] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea54311f-61f8-48fe-a951-42b406887df6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.549271] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1113.549271] env[62183]: value = "task-1387548" [ 1113.549271] env[62183]: _type = "Task" [ 1113.549271] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.556762] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387548, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.804131] env[62183]: DEBUG oslo_vmware.api [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387547, 'name': PowerOnVM_Task, 'duration_secs': 0.43876} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.804466] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1113.804684] env[62183]: INFO nova.compute.manager [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Took 7.14 seconds to spawn the instance on the hypervisor. [ 1113.804869] env[62183]: DEBUG nova.compute.manager [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1113.805707] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7c3ee1-d3f2-4602-a6ad-41ffbb06eefd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.925599] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387545, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.059537] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387548, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.322187] env[62183]: INFO nova.compute.manager [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Took 13.70 seconds to build instance. [ 1114.426326] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387545, 'name': ReconfigVM_Task, 'duration_secs': 1.477886} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.426731] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance '5aed3b2c-627c-4f33-8f22-950e758f822d' progress to 33 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1114.501809] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4c4b7a-9946-4fd7-b7f2-22bb4963dcb3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.509351] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f849cb92-4d1a-4868-9dfa-1c6fad8b50c8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.540370] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f81182-1921-459b-a658-a44a1dc2bbab {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.547883] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9a9222-34de-4634-821d-670390885c93 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.563606] env[62183]: DEBUG nova.compute.provider_tree [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.567574] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387548, 'name': ReconfigVM_Task, 'duration_secs': 0.563926} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.568054] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Reconfigured VM instance instance-00000065 to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1114.573306] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c55e2ad5-8c4c-4bc9-b093-252a152d863d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.589405] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1114.589405] env[62183]: value = "task-1387549" [ 1114.589405] env[62183]: _type = "Task" [ 1114.589405] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.597392] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387549, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.824255] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9302b97e-2c4f-48f1-8ccb-2e4e5a89354b tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.213s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.933055] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1114.933330] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1114.933494] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1114.933872] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1114.933872] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1114.933961] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1114.934179] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1114.934339] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1114.934549] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1114.934745] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1114.934931] env[62183]: DEBUG nova.virt.hardware [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1114.940244] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1114.940555] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23ab6aae-be18-4475-b636-6778c51ce8bc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.967352] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1114.967352] env[62183]: value = "task-1387550" [ 1114.967352] env[62183]: _type = "Task" [ 1114.967352] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.981824] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387550, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.002130] env[62183]: DEBUG nova.compute.manager [req-1e87131e-b94c-4022-90f4-a9e4b905fd14 req-e6cc8a60-2d1b-43a9-afa5-e7036a96f385 service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Received event network-changed-c7b92447-a61f-4d71-b39f-28af011a2142 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.002130] env[62183]: DEBUG nova.compute.manager [req-1e87131e-b94c-4022-90f4-a9e4b905fd14 req-e6cc8a60-2d1b-43a9-afa5-e7036a96f385 service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Refreshing instance network info cache due to event network-changed-c7b92447-a61f-4d71-b39f-28af011a2142. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1115.002398] env[62183]: DEBUG oslo_concurrency.lockutils [req-1e87131e-b94c-4022-90f4-a9e4b905fd14 req-e6cc8a60-2d1b-43a9-afa5-e7036a96f385 service nova] Acquiring lock "refresh_cache-632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.002548] env[62183]: DEBUG oslo_concurrency.lockutils [req-1e87131e-b94c-4022-90f4-a9e4b905fd14 req-e6cc8a60-2d1b-43a9-afa5-e7036a96f385 service nova] Acquired lock "refresh_cache-632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.002810] env[62183]: DEBUG nova.network.neutron [req-1e87131e-b94c-4022-90f4-a9e4b905fd14 req-e6cc8a60-2d1b-43a9-afa5-e7036a96f385 service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Refreshing network info cache for port c7b92447-a61f-4d71-b39f-28af011a2142 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1115.069289] env[62183]: DEBUG nova.scheduler.client.report [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1115.098235] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387549, 'name': ReconfigVM_Task, 'duration_secs': 0.124596} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.098543] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294524', 'volume_id': '6b106861-b013-4124-a712-16076f82c288', 'name': 'volume-6b106861-b013-4124-a712-16076f82c288', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '88846abe-9e6e-4a88-bf6e-9978610213da', 'attached_at': '2024-10-31T11:15:32.000000', 'detached_at': '', 'volume_id': '6b106861-b013-4124-a712-16076f82c288', 'serial': '6b106861-b013-4124-a712-16076f82c288'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1115.098839] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1115.099623] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbaeb79-3123-4612-abc8-cb0dcbb4d22a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.106244] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1115.106472] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1915ec32-560f-4459-9a36-8c5379a0775d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.189019] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1115.189273] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1115.189469] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleting the datastore file [datastore2] 88846abe-9e6e-4a88-bf6e-9978610213da {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1115.189749] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07e57bda-e795-406f-b146-97e8b74893aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.196724] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1115.196724] env[62183]: value = "task-1387552" [ 1115.196724] env[62183]: _type = "Task" [ 1115.196724] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.205833] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387552, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.478619] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387550, 'name': ReconfigVM_Task, 'duration_secs': 0.158128} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.478926] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1115.479641] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49b0f3a-b6de-43ea-99fd-8ff851aca1f9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.500977] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 5aed3b2c-627c-4f33-8f22-950e758f822d/5aed3b2c-627c-4f33-8f22-950e758f822d.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1115.501257] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-285ec3e3-da14-4e4e-9c86-2f14f76071ed {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.522045] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1115.522045] env[62183]: value = "task-1387553" [ 1115.522045] env[62183]: _type = "Task" [ 1115.522045] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.530495] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387553, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.574372] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.153s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.574916] env[62183]: DEBUG nova.compute.manager [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1115.707412] env[62183]: DEBUG oslo_vmware.api [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387552, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145794} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.707763] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1115.708030] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1115.708284] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1115.708543] env[62183]: INFO nova.compute.manager [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Took 2.77 seconds to destroy the instance on the hypervisor. [ 1115.708872] env[62183]: DEBUG oslo.service.loopingcall [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.711589] env[62183]: DEBUG nova.compute.manager [-] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1115.711716] env[62183]: DEBUG nova.network.neutron [-] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1115.782010] env[62183]: DEBUG nova.network.neutron [req-1e87131e-b94c-4022-90f4-a9e4b905fd14 req-e6cc8a60-2d1b-43a9-afa5-e7036a96f385 service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Updated VIF entry in instance network info cache for port c7b92447-a61f-4d71-b39f-28af011a2142. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1115.782400] env[62183]: DEBUG nova.network.neutron [req-1e87131e-b94c-4022-90f4-a9e4b905fd14 req-e6cc8a60-2d1b-43a9-afa5-e7036a96f385 service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Updating instance_info_cache with network_info: [{"id": "c7b92447-a61f-4d71-b39f-28af011a2142", "address": "fa:16:3e:f9:e5:a6", "network": {"id": "5926bcfe-2e34-49fd-bc94-64dc8e6720b1", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-135849370-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06e81bd1a81d4009ae2a75fe819f9b7c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "205fb402-8eaf-4b61-8f57-8f216024179a", "external-id": "nsx-vlan-transportzone-78", "segmentation_id": 78, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7b92447-a6", "ovs_interfaceid": "c7b92447-a61f-4d71-b39f-28af011a2142", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.031379] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387553, 'name': ReconfigVM_Task, 'duration_secs': 0.261419} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.031679] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 5aed3b2c-627c-4f33-8f22-950e758f822d/5aed3b2c-627c-4f33-8f22-950e758f822d.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1116.031943] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance '5aed3b2c-627c-4f33-8f22-950e758f822d' progress to 50 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1116.079785] env[62183]: DEBUG nova.compute.utils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1116.081498] env[62183]: DEBUG nova.compute.manager [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1116.081498] env[62183]: DEBUG nova.network.neutron [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1116.136334] env[62183]: DEBUG nova.policy [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbf2ac3d1b8f46cf89df0e7a106c75e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ffc6a4c3a1e74a9b8d9eae70dce238a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1116.261974] env[62183]: DEBUG nova.compute.manager [req-467bb8b6-d034-4c6a-9327-ba00676fdce9 req-3df1e034-6e66-47f7-be3b-0c9e610d2427 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Received event network-vif-deleted-be785bab-5120-4093-b24c-25e5c14608bf {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1116.262213] env[62183]: INFO nova.compute.manager [req-467bb8b6-d034-4c6a-9327-ba00676fdce9 req-3df1e034-6e66-47f7-be3b-0c9e610d2427 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Neutron deleted interface be785bab-5120-4093-b24c-25e5c14608bf; detaching it from the instance and deleting it from the info cache [ 1116.262395] env[62183]: DEBUG nova.network.neutron [req-467bb8b6-d034-4c6a-9327-ba00676fdce9 req-3df1e034-6e66-47f7-be3b-0c9e610d2427 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.285575] env[62183]: DEBUG oslo_concurrency.lockutils [req-1e87131e-b94c-4022-90f4-a9e4b905fd14 req-e6cc8a60-2d1b-43a9-afa5-e7036a96f385 service nova] Releasing lock "refresh_cache-632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.459944] env[62183]: DEBUG nova.network.neutron [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Successfully created port: 2de4249d-027d-4a0c-aab3-bc1828b18825 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1116.538585] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589ebd41-13af-4161-9ecc-bbf1727456e4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.558844] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b38ec604-d575-4ab8-b99d-116fee731816 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.576120] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance '5aed3b2c-627c-4f33-8f22-950e758f822d' progress to 67 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1116.587597] env[62183]: DEBUG nova.compute.manager [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1116.737465] env[62183]: DEBUG nova.network.neutron [-] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.765709] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f248bd73-50f4-40af-947c-7d6ed2fc5357 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.774893] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4249d9fd-eadf-4acb-a0f6-f33ed12b86ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.800328] env[62183]: DEBUG nova.compute.manager [req-467bb8b6-d034-4c6a-9327-ba00676fdce9 req-3df1e034-6e66-47f7-be3b-0c9e610d2427 service nova] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Detach interface failed, port_id=be785bab-5120-4093-b24c-25e5c14608bf, reason: Instance 88846abe-9e6e-4a88-bf6e-9978610213da could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1117.115042] env[62183]: DEBUG nova.network.neutron [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Port d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd binding to destination host cpu-1 is already ACTIVE {{(pid=62183) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1117.239843] env[62183]: INFO nova.compute.manager [-] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Took 1.53 seconds to deallocate network for instance. [ 1117.596382] env[62183]: DEBUG nova.compute.manager [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1117.624033] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1117.624283] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1117.624443] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1117.624626] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1117.624775] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1117.624924] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1117.625144] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1117.625310] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1117.625478] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1117.625643] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1117.625816] env[62183]: DEBUG nova.virt.hardware [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.626663] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b7e2b06-05a1-43e3-9c5d-10791005741e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.634866] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a4217c-0d57-41e7-b2a3-7720e40cb408 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.785739] env[62183]: INFO nova.compute.manager [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 88846abe-9e6e-4a88-bf6e-9978610213da] Took 0.55 seconds to detach 1 volumes for instance. [ 1118.005931] env[62183]: DEBUG nova.network.neutron [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Successfully updated port: 2de4249d-027d-4a0c-aab3-bc1828b18825 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1118.143796] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "5aed3b2c-627c-4f33-8f22-950e758f822d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.144036] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.144169] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.286272] env[62183]: DEBUG nova.compute.manager [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Received event network-vif-plugged-2de4249d-027d-4a0c-aab3-bc1828b18825 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1118.286482] env[62183]: DEBUG oslo_concurrency.lockutils [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.286679] env[62183]: DEBUG oslo_concurrency.lockutils [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.286855] env[62183]: DEBUG oslo_concurrency.lockutils [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.287040] env[62183]: DEBUG nova.compute.manager [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] No waiting events found dispatching network-vif-plugged-2de4249d-027d-4a0c-aab3-bc1828b18825 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1118.287212] env[62183]: WARNING nova.compute.manager [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Received unexpected event network-vif-plugged-2de4249d-027d-4a0c-aab3-bc1828b18825 for instance with vm_state building and task_state spawning. [ 1118.287427] env[62183]: DEBUG nova.compute.manager [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Received event network-changed-2de4249d-027d-4a0c-aab3-bc1828b18825 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1118.287610] env[62183]: DEBUG nova.compute.manager [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Refreshing instance network info cache due to event network-changed-2de4249d-027d-4a0c-aab3-bc1828b18825. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1118.287800] env[62183]: DEBUG oslo_concurrency.lockutils [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] Acquiring lock "refresh_cache-1b411fab-7728-4b29-ad9f-ce22a2face55" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.287939] env[62183]: DEBUG oslo_concurrency.lockutils [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] Acquired lock "refresh_cache-1b411fab-7728-4b29-ad9f-ce22a2face55" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.288111] env[62183]: DEBUG nova.network.neutron [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Refreshing network info cache for port 2de4249d-027d-4a0c-aab3-bc1828b18825 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1118.292296] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.292522] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.292709] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.312789] env[62183]: INFO nova.scheduler.client.report [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted allocations for instance 88846abe-9e6e-4a88-bf6e-9978610213da [ 1118.508282] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "refresh_cache-1b411fab-7728-4b29-ad9f-ce22a2face55" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.819476] env[62183]: DEBUG nova.network.neutron [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1118.823625] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b69159db-aa0d-4281-8928-7c6876478274 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "88846abe-9e6e-4a88-bf6e-9978610213da" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.887s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.898787] env[62183]: DEBUG nova.network.neutron [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.178445] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.178681] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.178907] env[62183]: DEBUG nova.network.neutron [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1119.401403] env[62183]: DEBUG oslo_concurrency.lockutils [req-8365b329-f2c4-44c8-b736-b51c2e26302e req-78e515ff-66b0-438a-91fb-2ceadda108e5 service nova] Releasing lock "refresh_cache-1b411fab-7728-4b29-ad9f-ce22a2face55" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.401806] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired lock "refresh_cache-1b411fab-7728-4b29-ad9f-ce22a2face55" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.401980] env[62183]: DEBUG nova.network.neutron [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1119.921790] env[62183]: DEBUG nova.network.neutron [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance_info_cache with network_info: [{"id": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "address": "fa:16:3e:95:fa:9f", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2e95f9f-6c", "ovs_interfaceid": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.933237] env[62183]: DEBUG nova.network.neutron [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1120.060037] env[62183]: DEBUG nova.network.neutron [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Updating instance_info_cache with network_info: [{"id": "2de4249d-027d-4a0c-aab3-bc1828b18825", "address": "fa:16:3e:94:55:d0", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2de4249d-02", "ovs_interfaceid": "2de4249d-027d-4a0c-aab3-bc1828b18825", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.083986] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.084296] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.425163] env[62183]: DEBUG oslo_concurrency.lockutils [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.562913] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Releasing lock "refresh_cache-1b411fab-7728-4b29-ad9f-ce22a2face55" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.563275] env[62183]: DEBUG nova.compute.manager [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Instance network_info: |[{"id": "2de4249d-027d-4a0c-aab3-bc1828b18825", "address": "fa:16:3e:94:55:d0", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2de4249d-02", "ovs_interfaceid": "2de4249d-027d-4a0c-aab3-bc1828b18825", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1120.563720] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:55:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dbdab640-5fea-4254-8bd3-f855b7eaca0d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2de4249d-027d-4a0c-aab3-bc1828b18825', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1120.571303] env[62183]: DEBUG oslo.service.loopingcall [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1120.571571] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1120.571860] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a525189-a88e-4bfb-9f63-7cbf7895e564 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.588227] env[62183]: DEBUG nova.compute.manager [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1120.592156] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1120.592156] env[62183]: value = "task-1387554" [ 1120.592156] env[62183]: _type = "Task" [ 1120.592156] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.600472] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387554, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.948763] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffadfdd-7a9c-4302-b996-afba20eecbe0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.967513] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695414cc-d054-4fde-a86c-a4e6c116966f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.974332] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance '5aed3b2c-627c-4f33-8f22-950e758f822d' progress to 83 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1121.106825] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387554, 'name': CreateVM_Task, 'duration_secs': 0.288075} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.106825] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1121.107120] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.107583] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.107630] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1121.107859] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42b02c53-a326-496f-b184-8604a3240e87 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.111028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.111028] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.112453] env[62183]: INFO nova.compute.claims [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1121.117042] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1121.117042] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52bf7e2f-04ee-df92-a154-63ca7c320b52" [ 1121.117042] env[62183]: _type = "Task" [ 1121.117042] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.123766] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bf7e2f-04ee-df92-a154-63ca7c320b52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.480778] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1121.481064] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9daef1b8-8b6e-483c-8a9d-23820276dc63 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.487982] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1121.487982] env[62183]: value = "task-1387555" [ 1121.487982] env[62183]: _type = "Task" [ 1121.487982] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.497180] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.628688] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bf7e2f-04ee-df92-a154-63ca7c320b52, 'name': SearchDatastore_Task, 'duration_secs': 0.019056} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.629560] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.629825] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1121.630097] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.630258] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.630451] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1121.630950] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c4a56e8-afd3-4a72-8f94-85711bd00a00 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.639016] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1121.639207] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1121.639873] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c04a6df9-9d31-4edc-9f5e-d2df896d7e88 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.644823] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1121.644823] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]528babf6-295d-5649-4a6f-b32d0699b29e" [ 1121.644823] env[62183]: _type = "Task" [ 1121.644823] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.652124] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528babf6-295d-5649-4a6f-b32d0699b29e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.997623] env[62183]: DEBUG oslo_vmware.api [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387555, 'name': PowerOnVM_Task, 'duration_secs': 0.474859} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.997986] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1121.998069] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-21320fe4-1482-4a07-8c2c-c1d172da5ec9 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance '5aed3b2c-627c-4f33-8f22-950e758f822d' progress to 100 {{(pid=62183) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1122.154825] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]528babf6-295d-5649-4a6f-b32d0699b29e, 'name': SearchDatastore_Task, 'duration_secs': 0.007518} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.155706] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ebdf36d-3760-4f9d-bf1f-d064728cae18 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.160407] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1122.160407] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52203079-a485-a5cc-3a62-e1b9988a142f" [ 1122.160407] env[62183]: _type = "Task" [ 1122.160407] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.169991] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52203079-a485-a5cc-3a62-e1b9988a142f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.192309] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e1149e-5ccb-4ac1-9c16-93d7b5dd5f9c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.198857] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562b1e06-3899-415b-b204-ae397cd28b0a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.227688] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7627bf-45fe-4b14-af71-38e2b74d51dd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.234805] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15d11f1-57b1-4ed4-ba6f-630116c6046b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.248912] env[62183]: DEBUG nova.compute.provider_tree [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.670791] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52203079-a485-a5cc-3a62-e1b9988a142f, 'name': SearchDatastore_Task, 'duration_secs': 0.008476} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.671080] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.671350] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 1b411fab-7728-4b29-ad9f-ce22a2face55/1b411fab-7728-4b29-ad9f-ce22a2face55.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1122.671607] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e88fb7f7-96f9-4639-806b-3ee31ea5e1bf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.678068] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1122.678068] env[62183]: value = "task-1387556" [ 1122.678068] env[62183]: _type = "Task" [ 1122.678068] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.685817] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.752141] env[62183]: DEBUG nova.scheduler.client.report [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1123.188340] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387556, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.258054] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.146s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.258054] env[62183]: DEBUG nova.compute.manager [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1123.688821] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387556, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.584105} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.689125] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 1b411fab-7728-4b29-ad9f-ce22a2face55/1b411fab-7728-4b29-ad9f-ce22a2face55.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1123.689350] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1123.689604] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a61da8e2-cae6-4eb2-b1d3-c85bc5124bfd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.696688] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1123.696688] env[62183]: value = "task-1387557" [ 1123.696688] env[62183]: _type = "Task" [ 1123.696688] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.704695] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.762844] env[62183]: DEBUG nova.compute.utils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1123.764344] env[62183]: DEBUG nova.compute.manager [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1123.764521] env[62183]: DEBUG nova.network.neutron [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1123.801821] env[62183]: DEBUG nova.policy [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4b92513f568466e81075af3fa4604fa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88022032e6e04a4f96bc49c2ca5ede29', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1124.058830] env[62183]: DEBUG nova.network.neutron [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Successfully created port: 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1124.206788] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.322327} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.207195] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1124.207816] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2379d9b-3761-4533-aa79-4c484730d23f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.229221] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 1b411fab-7728-4b29-ad9f-ce22a2face55/1b411fab-7728-4b29-ad9f-ce22a2face55.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.229992] env[62183]: DEBUG nova.network.neutron [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Port d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd binding to destination host cpu-1 is already ACTIVE {{(pid=62183) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1124.230244] env[62183]: DEBUG oslo_concurrency.lockutils [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.230404] env[62183]: DEBUG oslo_concurrency.lockutils [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.230568] env[62183]: DEBUG nova.network.neutron [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1124.231657] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3d8b7f2-8840-4e28-8215-2f3f1dbae79f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.253828] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1124.253828] env[62183]: value = "task-1387558" [ 1124.253828] env[62183]: _type = "Task" [ 1124.253828] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.262067] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387558, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.266951] env[62183]: DEBUG nova.compute.manager [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1124.535775] env[62183]: DEBUG nova.network.neutron [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance_info_cache with network_info: [{"id": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "address": "fa:16:3e:95:fa:9f", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2e95f9f-6c", "ovs_interfaceid": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.764434] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387558, 'name': ReconfigVM_Task, 'duration_secs': 0.387959} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.764746] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 1b411fab-7728-4b29-ad9f-ce22a2face55/1b411fab-7728-4b29-ad9f-ce22a2face55.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1124.765398] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fba4cd55-c641-4989-b9e7-43cd2a107fd1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.775790] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1124.775790] env[62183]: value = "task-1387559" [ 1124.775790] env[62183]: _type = "Task" [ 1124.775790] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.783927] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387559, 'name': Rename_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.039271] env[62183]: DEBUG oslo_concurrency.lockutils [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.277263] env[62183]: DEBUG nova.compute.manager [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1125.288517] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387559, 'name': Rename_Task} progress is 14%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.301999] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1125.302256] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1125.302418] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1125.302643] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1125.302867] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1125.303045] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1125.303267] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1125.303430] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1125.303602] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1125.303767] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1125.303939] env[62183]: DEBUG nova.virt.hardware [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1125.304756] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f51ace0-1a02-4940-b573-9c3633392558 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.312477] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ee633e-d374-4f31-9465-3be867ec8741 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.542863] env[62183]: DEBUG nova.compute.manager [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62183) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1125.543131] env[62183]: DEBUG oslo_concurrency.lockutils [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.543456] env[62183]: DEBUG oslo_concurrency.lockutils [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.590701] env[62183]: DEBUG nova.compute.manager [req-2d36f352-2096-49f1-b51c-7f113910d577 req-3f10cc9b-a6d8-41c3-9815-3817c4d515da service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received event network-vif-plugged-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1125.590932] env[62183]: DEBUG oslo_concurrency.lockutils [req-2d36f352-2096-49f1-b51c-7f113910d577 req-3f10cc9b-a6d8-41c3-9815-3817c4d515da service nova] Acquiring lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.591173] env[62183]: DEBUG oslo_concurrency.lockutils [req-2d36f352-2096-49f1-b51c-7f113910d577 req-3f10cc9b-a6d8-41c3-9815-3817c4d515da service nova] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.591423] env[62183]: DEBUG oslo_concurrency.lockutils [req-2d36f352-2096-49f1-b51c-7f113910d577 req-3f10cc9b-a6d8-41c3-9815-3817c4d515da service nova] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.591612] env[62183]: DEBUG nova.compute.manager [req-2d36f352-2096-49f1-b51c-7f113910d577 req-3f10cc9b-a6d8-41c3-9815-3817c4d515da service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] No waiting events found dispatching network-vif-plugged-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1125.591660] env[62183]: WARNING nova.compute.manager [req-2d36f352-2096-49f1-b51c-7f113910d577 req-3f10cc9b-a6d8-41c3-9815-3817c4d515da service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received unexpected event network-vif-plugged-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a for instance with vm_state building and task_state spawning. [ 1125.678746] env[62183]: DEBUG nova.network.neutron [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Successfully updated port: 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1125.789238] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387559, 'name': Rename_Task, 'duration_secs': 0.93532} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.789497] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1125.789793] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4bb593e0-64e6-4b14-b3df-789714a0aaf9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.795608] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1125.795608] env[62183]: value = "task-1387560" [ 1125.795608] env[62183]: _type = "Task" [ 1125.795608] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.802686] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387560, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.045949] env[62183]: DEBUG nova.objects.instance [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'migration_context' on Instance uuid 5aed3b2c-627c-4f33-8f22-950e758f822d {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.185811] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.186034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.186118] env[62183]: DEBUG nova.network.neutron [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1126.310613] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387560, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.622039] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed27259e-3370-49c0-b01c-137539f7cfeb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.629374] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3079b0be-7d12-4185-aa22-75a8df397e4c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.658441] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabb4dab-6d95-4144-bb25-871226b25c50 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.665205] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0f2ce7-504d-4ea7-bf30-9a4bf9a00ca3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.679200] env[62183]: DEBUG nova.compute.provider_tree [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1126.717285] env[62183]: DEBUG nova.network.neutron [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1126.805390] env[62183]: DEBUG oslo_vmware.api [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387560, 'name': PowerOnVM_Task, 'duration_secs': 0.530841} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.807481] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1126.807695] env[62183]: INFO nova.compute.manager [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Took 9.21 seconds to spawn the instance on the hypervisor. [ 1126.807878] env[62183]: DEBUG nova.compute.manager [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1126.808655] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135e98f0-814f-4a9c-9fd6-7bae14eac724 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.843150] env[62183]: DEBUG nova.network.neutron [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating instance_info_cache with network_info: [{"id": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "address": "fa:16:3e:80:88:09", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ebdb1ac-70", "ovs_interfaceid": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.199369] env[62183]: ERROR nova.scheduler.client.report [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [req-179cf23a-b888-40d6-a2b3-9cd2a26770fa] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-179cf23a-b888-40d6-a2b3-9cd2a26770fa"}]} [ 1127.214884] env[62183]: DEBUG nova.scheduler.client.report [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1127.226585] env[62183]: DEBUG nova.scheduler.client.report [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1127.226807] env[62183]: DEBUG nova.compute.provider_tree [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1127.238419] env[62183]: DEBUG nova.scheduler.client.report [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1127.255819] env[62183]: DEBUG nova.scheduler.client.report [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1127.319018] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6a3c4c-32d1-4f07-be53-b8ac5938a19a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.335388] env[62183]: INFO nova.compute.manager [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Took 13.93 seconds to build instance. [ 1127.338715] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064e6c46-86a9-4b8a-9fc6-adfbf283680b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.344975] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.345807] env[62183]: DEBUG nova.compute.manager [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Instance network_info: |[{"id": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "address": "fa:16:3e:80:88:09", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ebdb1ac-70", "ovs_interfaceid": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1127.346456] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:88:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ff90ec9-3c7e-4e76-b409-fcf37fc588d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ebdb1ac-70ed-4649-95b0-4d71bd7be73a', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1127.355257] env[62183]: DEBUG oslo.service.loopingcall [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.379758] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1127.380415] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-973b4e65-7b79-425d-8319-eb8b1ef6ea1b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.394505] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efeaef5f-6a51-4961-8e9f-8cc3709e5a80 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.403409] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84f2b34-4151-4b11-8526-0ac577ea73be {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.407362] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1127.407362] env[62183]: value = "task-1387561" [ 1127.407362] env[62183]: _type = "Task" [ 1127.407362] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.418060] env[62183]: DEBUG nova.compute.provider_tree [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1127.424345] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387561, 'name': CreateVM_Task} progress is 10%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.623658] env[62183]: DEBUG nova.compute.manager [req-f618db66-d08f-4214-bb49-4de19133bf18 req-2ebdce24-5abf-4a4c-a136-e64c36472855 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received event network-changed-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1127.623873] env[62183]: DEBUG nova.compute.manager [req-f618db66-d08f-4214-bb49-4de19133bf18 req-2ebdce24-5abf-4a4c-a136-e64c36472855 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Refreshing instance network info cache due to event network-changed-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1127.624113] env[62183]: DEBUG oslo_concurrency.lockutils [req-f618db66-d08f-4214-bb49-4de19133bf18 req-2ebdce24-5abf-4a4c-a136-e64c36472855 service nova] Acquiring lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.624267] env[62183]: DEBUG oslo_concurrency.lockutils [req-f618db66-d08f-4214-bb49-4de19133bf18 req-2ebdce24-5abf-4a4c-a136-e64c36472855 service nova] Acquired lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.624430] env[62183]: DEBUG nova.network.neutron [req-f618db66-d08f-4214-bb49-4de19133bf18 req-2ebdce24-5abf-4a4c-a136-e64c36472855 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Refreshing network info cache for port 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1127.844848] env[62183]: DEBUG oslo_concurrency.lockutils [None req-36d5b50e-0b04-496d-b45d-25bd6558812d tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.462s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.917280] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387561, 'name': CreateVM_Task, 'duration_secs': 0.28821} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.917448] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1127.918061] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.918248] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.918567] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1127.918820] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d43c6869-97c4-4ac4-a3ba-600fdf59830b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.926743] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1127.926743] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]521b471f-649e-e02a-4284-b50f960dec60" [ 1127.926743] env[62183]: _type = "Task" [ 1127.926743] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.934280] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521b471f-649e-e02a-4284-b50f960dec60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.949427] env[62183]: DEBUG nova.scheduler.client.report [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 143 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1127.949688] env[62183]: DEBUG nova.compute.provider_tree [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 143 to 144 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1127.949906] env[62183]: DEBUG nova.compute.provider_tree [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1128.320986] env[62183]: DEBUG nova.network.neutron [req-f618db66-d08f-4214-bb49-4de19133bf18 req-2ebdce24-5abf-4a4c-a136-e64c36472855 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updated VIF entry in instance network info cache for port 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1128.321623] env[62183]: DEBUG nova.network.neutron [req-f618db66-d08f-4214-bb49-4de19133bf18 req-2ebdce24-5abf-4a4c-a136-e64c36472855 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating instance_info_cache with network_info: [{"id": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "address": "fa:16:3e:80:88:09", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ebdb1ac-70", "ovs_interfaceid": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.438058] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]521b471f-649e-e02a-4284-b50f960dec60, 'name': SearchDatastore_Task, 'duration_secs': 0.022802} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.438058] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.438058] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1128.438058] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.438328] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.438328] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.438579] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34449f05-5478-4261-a3b4-cc798cb13505 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.449549] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.449729] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1128.450406] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acf347fc-c90b-4b2b-8932-466f3c72b45f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.458464] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1128.458464] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52674bda-fb9d-4852-fe75-d6044be2c63a" [ 1128.458464] env[62183]: _type = "Task" [ 1128.458464] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.465757] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52674bda-fb9d-4852-fe75-d6044be2c63a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.825400] env[62183]: DEBUG oslo_concurrency.lockutils [req-f618db66-d08f-4214-bb49-4de19133bf18 req-2ebdce24-5abf-4a4c-a136-e64c36472855 service nova] Releasing lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.961764] env[62183]: DEBUG oslo_concurrency.lockutils [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.418s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.982578] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52674bda-fb9d-4852-fe75-d6044be2c63a, 'name': SearchDatastore_Task, 'duration_secs': 0.046767} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.983396] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9edc78a9-a7a2-4067-86ae-e444a60b7de1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.990215] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1128.990215] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f33ce7-5879-1835-0466-2ab9b643ab06" [ 1128.990215] env[62183]: _type = "Task" [ 1128.990215] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.998728] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f33ce7-5879-1835-0466-2ab9b643ab06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.501208] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f33ce7-5879-1835-0466-2ab9b643ab06, 'name': SearchDatastore_Task, 'duration_secs': 0.010255} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.501584] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "[datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.501746] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9/2d6f015c-5189-4a0b-bb79-3180d557e8f9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1129.502034] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac4f5796-fdc3-4333-8629-40632e66f2b5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.508760] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1129.508760] env[62183]: value = "task-1387562" [ 1129.508760] env[62183]: _type = "Task" [ 1129.508760] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.516767] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387562, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.674330] env[62183]: DEBUG nova.compute.manager [req-d08eebc0-bd1b-4900-bed4-aa08046333bb req-0c1e3bcb-606e-4734-b0ef-c0a2aa2af1a1 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Received event network-changed-2de4249d-027d-4a0c-aab3-bc1828b18825 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.674553] env[62183]: DEBUG nova.compute.manager [req-d08eebc0-bd1b-4900-bed4-aa08046333bb req-0c1e3bcb-606e-4734-b0ef-c0a2aa2af1a1 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Refreshing instance network info cache due to event network-changed-2de4249d-027d-4a0c-aab3-bc1828b18825. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1129.674788] env[62183]: DEBUG oslo_concurrency.lockutils [req-d08eebc0-bd1b-4900-bed4-aa08046333bb req-0c1e3bcb-606e-4734-b0ef-c0a2aa2af1a1 service nova] Acquiring lock "refresh_cache-1b411fab-7728-4b29-ad9f-ce22a2face55" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.674940] env[62183]: DEBUG oslo_concurrency.lockutils [req-d08eebc0-bd1b-4900-bed4-aa08046333bb req-0c1e3bcb-606e-4734-b0ef-c0a2aa2af1a1 service nova] Acquired lock "refresh_cache-1b411fab-7728-4b29-ad9f-ce22a2face55" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.675134] env[62183]: DEBUG nova.network.neutron [req-d08eebc0-bd1b-4900-bed4-aa08046333bb req-0c1e3bcb-606e-4734-b0ef-c0a2aa2af1a1 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Refreshing network info cache for port 2de4249d-027d-4a0c-aab3-bc1828b18825 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1130.020027] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387562, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.432349] env[62183]: DEBUG nova.network.neutron [req-d08eebc0-bd1b-4900-bed4-aa08046333bb req-0c1e3bcb-606e-4734-b0ef-c0a2aa2af1a1 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Updated VIF entry in instance network info cache for port 2de4249d-027d-4a0c-aab3-bc1828b18825. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1130.432728] env[62183]: DEBUG nova.network.neutron [req-d08eebc0-bd1b-4900-bed4-aa08046333bb req-0c1e3bcb-606e-4734-b0ef-c0a2aa2af1a1 service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Updating instance_info_cache with network_info: [{"id": "2de4249d-027d-4a0c-aab3-bc1828b18825", "address": "fa:16:3e:94:55:d0", "network": {"id": "a9472080-1a71-4608-848f-3101fb8466c2", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-745824666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.223", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffc6a4c3a1e74a9b8d9eae70dce238a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dbdab640-5fea-4254-8bd3-f855b7eaca0d", "external-id": "nsx-vlan-transportzone-615", "segmentation_id": 615, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2de4249d-02", "ovs_interfaceid": "2de4249d-027d-4a0c-aab3-bc1828b18825", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.504511] env[62183]: INFO nova.compute.manager [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Swapping old allocation on dict_keys(['09c07e5d-2ed9-41c2-be62-db0f731d0b87']) held by migration 27011ab2-e2da-49dc-967a-0a676032d5ed for instance [ 1130.520250] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387562, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641321} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.522735] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9/2d6f015c-5189-4a0b-bb79-3180d557e8f9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1130.522935] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1130.523384] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88378ca6-51a6-4ea4-861c-4b95d7170125 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.530088] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1130.530088] env[62183]: value = "task-1387563" [ 1130.530088] env[62183]: _type = "Task" [ 1130.530088] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.534326] env[62183]: DEBUG nova.scheduler.client.report [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Overwriting current allocation {'allocations': {'09c07e5d-2ed9-41c2-be62-db0f731d0b87': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 144}}, 'project_id': '34581a9b5c1943eabc13bbb300a0f086', 'user_id': 'b85b5dce9bf44da6afd11f100a121f03', 'consumer_generation': 1} on consumer 5aed3b2c-627c-4f33-8f22-950e758f822d {{(pid=62183) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2033}} [ 1130.538972] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387563, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.619721] env[62183]: DEBUG oslo_concurrency.lockutils [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.619965] env[62183]: DEBUG oslo_concurrency.lockutils [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.620115] env[62183]: DEBUG nova.network.neutron [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1130.936068] env[62183]: DEBUG oslo_concurrency.lockutils [req-d08eebc0-bd1b-4900-bed4-aa08046333bb req-0c1e3bcb-606e-4734-b0ef-c0a2aa2af1a1 service nova] Releasing lock "refresh_cache-1b411fab-7728-4b29-ad9f-ce22a2face55" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.039311] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387563, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080255} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.039586] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1131.040381] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7209020c-7335-4415-95fc-65724d7a7a81 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.061584] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9/2d6f015c-5189-4a0b-bb79-3180d557e8f9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1131.061811] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a001c87-ee49-451d-83fd-70c402fbb199 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.080636] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1131.080636] env[62183]: value = "task-1387564" [ 1131.080636] env[62183]: _type = "Task" [ 1131.080636] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.087955] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387564, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.319903] env[62183]: DEBUG nova.network.neutron [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance_info_cache with network_info: [{"id": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "address": "fa:16:3e:95:fa:9f", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2e95f9f-6c", "ovs_interfaceid": "d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.592176] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387564, 'name': ReconfigVM_Task, 'duration_secs': 0.384521} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.592546] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9/2d6f015c-5189-4a0b-bb79-3180d557e8f9.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1131.593104] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ed2698c-7f72-4915-87f3-316fcb1d1a64 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.599591] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1131.599591] env[62183]: value = "task-1387565" [ 1131.599591] env[62183]: _type = "Task" [ 1131.599591] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.608451] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387565, 'name': Rename_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.823200] env[62183]: DEBUG oslo_concurrency.lockutils [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-5aed3b2c-627c-4f33-8f22-950e758f822d" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.823711] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1131.824015] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ce98f1d-dedb-45f4-baff-bc2341c741f5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.831202] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1131.831202] env[62183]: value = "task-1387566" [ 1131.831202] env[62183]: _type = "Task" [ 1131.831202] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.838831] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387566, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.109622] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387565, 'name': Rename_Task, 'duration_secs': 0.148076} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.109929] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1132.110192] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a6858c0-5187-4b82-8810-ba016257692c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.115730] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1132.115730] env[62183]: value = "task-1387567" [ 1132.115730] env[62183]: _type = "Task" [ 1132.115730] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.122689] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387567, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.340434] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387566, 'name': PowerOffVM_Task, 'duration_secs': 0.212211} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.340709] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1132.341380] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1132.341598] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1132.341759] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1132.341947] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1132.342113] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1132.342270] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1132.342476] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1132.342641] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1132.342808] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1132.342971] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1132.343162] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1132.347982] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2693073-7a05-4ad0-96f8-768764464b4c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.363664] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1132.363664] env[62183]: value = "task-1387568" [ 1132.363664] env[62183]: _type = "Task" [ 1132.363664] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.371130] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387568, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.625923] env[62183]: DEBUG oslo_vmware.api [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387567, 'name': PowerOnVM_Task, 'duration_secs': 0.464144} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.626272] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1132.626418] env[62183]: INFO nova.compute.manager [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Took 7.35 seconds to spawn the instance on the hypervisor. [ 1132.626607] env[62183]: DEBUG nova.compute.manager [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1132.627399] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303f30dc-d2f3-4a46-9135-c2bc0be7971b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.873753] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387568, 'name': ReconfigVM_Task, 'duration_secs': 0.13357} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.874549] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267f76d2-4cd7-4ce7-8e60-60d915a64f87 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.898586] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1132.898843] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1132.899017] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1132.899215] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1132.899374] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1132.899530] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1132.899743] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1132.899910] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1132.900096] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1132.900270] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1132.900470] env[62183]: DEBUG nova.virt.hardware [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1132.901370] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ecebbd3-6b95-4dd5-a40e-cdd5540fabc0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.907329] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1132.907329] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52612185-53e9-c78f-02ad-72ee90b691df" [ 1132.907329] env[62183]: _type = "Task" [ 1132.907329] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.915514] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52612185-53e9-c78f-02ad-72ee90b691df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.146212] env[62183]: INFO nova.compute.manager [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Took 12.05 seconds to build instance. [ 1133.419575] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52612185-53e9-c78f-02ad-72ee90b691df, 'name': SearchDatastore_Task, 'duration_secs': 0.007499} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.424824] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1133.425129] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87f250b5-6849-41bb-aac1-792b433ba222 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.443470] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1133.443470] env[62183]: value = "task-1387569" [ 1133.443470] env[62183]: _type = "Task" [ 1133.443470] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.453015] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387569, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.649256] env[62183]: DEBUG oslo_concurrency.lockutils [None req-43b0abd6-e782-4ea3-8ebe-036a2d232a92 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.565s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.955309] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387569, 'name': ReconfigVM_Task, 'duration_secs': 0.192017} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.955309] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1133.955309] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe947f1-a8d0-4c84-9267-43c1f77fa0e4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.977076] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 5aed3b2c-627c-4f33-8f22-950e758f822d/5aed3b2c-627c-4f33-8f22-950e758f822d.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.977390] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f327601-bc39-4ff9-9c48-ca90c7ea89fc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.995919] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1133.995919] env[62183]: value = "task-1387570" [ 1133.995919] env[62183]: _type = "Task" [ 1133.995919] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.003894] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387570, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.264516] env[62183]: DEBUG nova.compute.manager [req-7475abcf-8b75-4ba7-aee4-8a8a863f1f29 req-9d91f748-fe1e-462c-8321-eebb4635ed47 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received event network-changed-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1134.264728] env[62183]: DEBUG nova.compute.manager [req-7475abcf-8b75-4ba7-aee4-8a8a863f1f29 req-9d91f748-fe1e-462c-8321-eebb4635ed47 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Refreshing instance network info cache due to event network-changed-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1134.264953] env[62183]: DEBUG oslo_concurrency.lockutils [req-7475abcf-8b75-4ba7-aee4-8a8a863f1f29 req-9d91f748-fe1e-462c-8321-eebb4635ed47 service nova] Acquiring lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.265120] env[62183]: DEBUG oslo_concurrency.lockutils [req-7475abcf-8b75-4ba7-aee4-8a8a863f1f29 req-9d91f748-fe1e-462c-8321-eebb4635ed47 service nova] Acquired lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.265291] env[62183]: DEBUG nova.network.neutron [req-7475abcf-8b75-4ba7-aee4-8a8a863f1f29 req-9d91f748-fe1e-462c-8321-eebb4635ed47 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Refreshing network info cache for port 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1134.506149] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387570, 'name': ReconfigVM_Task, 'duration_secs': 0.272921} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.506438] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 5aed3b2c-627c-4f33-8f22-950e758f822d/5aed3b2c-627c-4f33-8f22-950e758f822d.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.507260] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f46138-db20-4c4a-9024-e4aaf0dbc123 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.524566] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62410704-e4ae-4e08-a4b4-29ea2d211efc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.542335] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aad9747-fdf3-402e-abf7-a575d169514f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.561285] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ff62f5-979e-409e-8cef-ce401117b0c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.568445] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1134.568699] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c154eb1-22dd-48c1-8965-8fac4296a86a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.574774] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1134.574774] env[62183]: value = "task-1387571" [ 1134.574774] env[62183]: _type = "Task" [ 1134.574774] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.581902] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387571, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.985839] env[62183]: DEBUG nova.network.neutron [req-7475abcf-8b75-4ba7-aee4-8a8a863f1f29 req-9d91f748-fe1e-462c-8321-eebb4635ed47 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updated VIF entry in instance network info cache for port 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1134.986245] env[62183]: DEBUG nova.network.neutron [req-7475abcf-8b75-4ba7-aee4-8a8a863f1f29 req-9d91f748-fe1e-462c-8321-eebb4635ed47 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating instance_info_cache with network_info: [{"id": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "address": "fa:16:3e:80:88:09", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ebdb1ac-70", "ovs_interfaceid": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.085250] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387571, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.489541] env[62183]: DEBUG oslo_concurrency.lockutils [req-7475abcf-8b75-4ba7-aee4-8a8a863f1f29 req-9d91f748-fe1e-462c-8321-eebb4635ed47 service nova] Releasing lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.585111] env[62183]: DEBUG oslo_vmware.api [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387571, 'name': PowerOnVM_Task, 'duration_secs': 0.580094} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.585388] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.596555] env[62183]: INFO nova.compute.manager [None req-272c6bd7-7c95-4a5c-b7f9-0d125fe85aba tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance to original state: 'active' [ 1137.841467] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "5aed3b2c-627c-4f33-8f22-950e758f822d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.841791] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.842000] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "5aed3b2c-627c-4f33-8f22-950e758f822d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.842220] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.842391] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.844874] env[62183]: INFO nova.compute.manager [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Terminating instance [ 1137.847254] env[62183]: DEBUG nova.compute.manager [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1137.847497] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1137.848361] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ba7504-442e-4ee5-b134-77e0c6717156 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.856905] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.857150] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f01bdc6a-6566-416c-9aa2-3b8cbd354295 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.863308] env[62183]: DEBUG oslo_vmware.api [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1137.863308] env[62183]: value = "task-1387572" [ 1137.863308] env[62183]: _type = "Task" [ 1137.863308] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.871821] env[62183]: DEBUG oslo_vmware.api [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.373153] env[62183]: DEBUG oslo_vmware.api [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387572, 'name': PowerOffVM_Task, 'duration_secs': 0.181098} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.373782] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1138.373782] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1138.373914] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-01293758-da93-44cb-a214-2d0f4396c214 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.438346] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1138.438608] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1138.438809] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleting the datastore file [datastore2] 5aed3b2c-627c-4f33-8f22-950e758f822d {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.439105] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd4708e1-2a28-40c8-ac46-761c91688bd7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.445260] env[62183]: DEBUG oslo_vmware.api [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1138.445260] env[62183]: value = "task-1387574" [ 1138.445260] env[62183]: _type = "Task" [ 1138.445260] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.452658] env[62183]: DEBUG oslo_vmware.api [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387574, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.954707] env[62183]: DEBUG oslo_vmware.api [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387574, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192783} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.955060] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.955173] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1138.955357] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1138.955546] env[62183]: INFO nova.compute.manager [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1138.955791] env[62183]: DEBUG oslo.service.loopingcall [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.955987] env[62183]: DEBUG nova.compute.manager [-] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1138.956097] env[62183]: DEBUG nova.network.neutron [-] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1139.362055] env[62183]: DEBUG nova.compute.manager [req-1f1d5a92-f174-45c4-9c4c-23bdfa1e8a23 req-4d5ebff3-4c51-47db-97bf-7333cfc1f2df service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Received event network-vif-deleted-d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.362055] env[62183]: INFO nova.compute.manager [req-1f1d5a92-f174-45c4-9c4c-23bdfa1e8a23 req-4d5ebff3-4c51-47db-97bf-7333cfc1f2df service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Neutron deleted interface d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd; detaching it from the instance and deleting it from the info cache [ 1139.362231] env[62183]: DEBUG nova.network.neutron [req-1f1d5a92-f174-45c4-9c4c-23bdfa1e8a23 req-4d5ebff3-4c51-47db-97bf-7333cfc1f2df service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.835961] env[62183]: DEBUG nova.network.neutron [-] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.864370] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9cde02eb-60c4-49a3-bba1-4388bb43e1ff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.876201] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10e8419-669c-43bb-9cf9-6110e35d579a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.901794] env[62183]: DEBUG nova.compute.manager [req-1f1d5a92-f174-45c4-9c4c-23bdfa1e8a23 req-4d5ebff3-4c51-47db-97bf-7333cfc1f2df service nova] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Detach interface failed, port_id=d2e95f9f-6cee-45a9-80f2-3386ed4ca3bd, reason: Instance 5aed3b2c-627c-4f33-8f22-950e758f822d could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1140.338737] env[62183]: INFO nova.compute.manager [-] [instance: 5aed3b2c-627c-4f33-8f22-950e758f822d] Took 1.38 seconds to deallocate network for instance. [ 1140.846017] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.846341] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.846542] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.868248] env[62183]: INFO nova.scheduler.client.report [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted allocations for instance 5aed3b2c-627c-4f33-8f22-950e758f822d [ 1140.914316] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.914535] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.914682] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1141.375701] env[62183]: DEBUG oslo_concurrency.lockutils [None req-17860356-39a5-4dd6-a3d4-390751911070 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "5aed3b2c-627c-4f33-8f22-950e758f822d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.534s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.422108] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Didn't find any instances for network info cache update. {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1142.422108] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.422108] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.422108] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.422545] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.422545] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.422545] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.720660] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.720887] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.926265] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Getting list of instances from cluster (obj){ [ 1142.926265] env[62183]: value = "domain-c8" [ 1142.926265] env[62183]: _type = "ClusterComputeResource" [ 1142.926265] env[62183]: } {{(pid=62183) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1142.927365] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b47912-4525-48a0-8ac0-91a3647c501f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.940484] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Got total of 3 instances {{(pid=62183) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1142.940654] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Triggering sync for uuid 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11 {{(pid=62183) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1142.940841] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Triggering sync for uuid 1b411fab-7728-4b29-ad9f-ce22a2face55 {{(pid=62183) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1142.941012] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Triggering sync for uuid 2d6f015c-5189-4a0b-bb79-3180d557e8f9 {{(pid=62183) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1142.941359] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.941577] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.941841] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.942043] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.942283] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.942473] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.942660] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.942795] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1142.943494] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca0f095-08cd-4a13-822b-1a3d034751d4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.946285] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77bf4c50-d06e-4c74-a838-d28a966e69f3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.949025] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a857227-f773-4d20-bc6d-bba2edfec210 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.951077] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.222924] env[62183]: DEBUG nova.compute.manager [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1143.454790] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.455212] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.455212] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.455368] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1143.456286] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cf3adf-406d-48b5-9c69-1f7316502053 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.461691] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.520s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.464058] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.464406] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.522s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.465657] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5309ff1-78c1-4087-8061-5c762b504810 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.480349] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed15fa7-1a55-4974-8dd0-8a926b6e558a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.486673] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d241ff1-59e3-42e9-952d-8211d2aca333 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.516017] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180711MB free_disk=176GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1143.516172] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.516376] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.743967] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.538216] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1144.538464] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 1b411fab-7728-4b29-ad9f-ce22a2face55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1144.538523] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 2d6f015c-5189-4a0b-bb79-3180d557e8f9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.042753] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 87961eb2-d977-48c1-9eaf-c371eac9bbb1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1145.042998] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1145.043127] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1145.094504] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af721b8a-0945-4b70-8208-d07d80086995 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.101633] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da10ce67-a34f-455b-845a-762bea3ec8eb {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.131265] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b784ed4-2a6d-4ec6-ad0f-6b9a8fa2483f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.137996] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c0e9b5-547c-411d-ab23-053e5945c95f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.150485] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1145.669066] env[62183]: ERROR nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [req-0d2d87f9-2f39-4af0-b8a7-97f72d71c4fd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0d2d87f9-2f39-4af0-b8a7-97f72d71c4fd"}]} [ 1145.684491] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1145.696019] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1145.696198] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1145.705539] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1145.721402] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1145.770231] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2776bb48-d05b-4064-9f45-865bff118095 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.777720] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07d9fa3-5ef9-4e8a-bd66-1952e67946fc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.806237] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255c7db2-42c0-4209-a8aa-8ae123ae4ea8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.812958] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a080ec68-9587-4a2e-a96a-e9744d0ea652 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.825253] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1146.356094] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 146 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1146.356334] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 146 to 147 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1146.356484] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1146.861295] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1146.861674] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.345s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.861854] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.118s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.863515] env[62183]: INFO nova.compute.claims [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1147.926027] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f05850-e2a9-4da8-aa09-45d61fa9beef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.933731] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd478c61-254c-4561-8ce1-eb652d9b6103 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.963224] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b661de3-91aa-4e53-9eaf-e1c7d059ab5b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.970133] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec38237-fbae-4930-be70-7a9308ea3f87 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.982598] env[62183]: DEBUG nova.compute.provider_tree [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.486177] env[62183]: DEBUG nova.scheduler.client.report [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1148.991454] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.129s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.992055] env[62183]: DEBUG nova.compute.manager [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1149.497196] env[62183]: DEBUG nova.compute.utils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1149.498653] env[62183]: DEBUG nova.compute.manager [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1149.498836] env[62183]: DEBUG nova.network.neutron [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1149.552267] env[62183]: DEBUG nova.policy [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b85b5dce9bf44da6afd11f100a121f03', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '34581a9b5c1943eabc13bbb300a0f086', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1149.802945] env[62183]: DEBUG nova.network.neutron [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Successfully created port: ae7b0413-aa72-4f0c-ab0b-f5c62d874948 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1150.001990] env[62183]: DEBUG nova.compute.manager [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1151.013060] env[62183]: DEBUG nova.compute.manager [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1151.038854] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1151.039145] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1151.039310] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1151.039543] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1151.039739] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1151.039905] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1151.040136] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1151.040317] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1151.040493] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1151.040713] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1151.040900] env[62183]: DEBUG nova.virt.hardware [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1151.041807] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120a9e9b-e95d-4b9a-a4f2-7242cb16a89f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.049990] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b5df6a-653c-4a8f-889b-b729d68590ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.171909] env[62183]: DEBUG nova.compute.manager [req-19c036b1-1b48-4ecb-86ad-50847c6bf114 req-fe1ae6a2-242e-42b4-8bc8-0f8b218e6db4 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Received event network-vif-plugged-ae7b0413-aa72-4f0c-ab0b-f5c62d874948 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1151.172163] env[62183]: DEBUG oslo_concurrency.lockutils [req-19c036b1-1b48-4ecb-86ad-50847c6bf114 req-fe1ae6a2-242e-42b4-8bc8-0f8b218e6db4 service nova] Acquiring lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.172390] env[62183]: DEBUG oslo_concurrency.lockutils [req-19c036b1-1b48-4ecb-86ad-50847c6bf114 req-fe1ae6a2-242e-42b4-8bc8-0f8b218e6db4 service nova] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.172564] env[62183]: DEBUG oslo_concurrency.lockutils [req-19c036b1-1b48-4ecb-86ad-50847c6bf114 req-fe1ae6a2-242e-42b4-8bc8-0f8b218e6db4 service nova] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.172808] env[62183]: DEBUG nova.compute.manager [req-19c036b1-1b48-4ecb-86ad-50847c6bf114 req-fe1ae6a2-242e-42b4-8bc8-0f8b218e6db4 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] No waiting events found dispatching network-vif-plugged-ae7b0413-aa72-4f0c-ab0b-f5c62d874948 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1151.172918] env[62183]: WARNING nova.compute.manager [req-19c036b1-1b48-4ecb-86ad-50847c6bf114 req-fe1ae6a2-242e-42b4-8bc8-0f8b218e6db4 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Received unexpected event network-vif-plugged-ae7b0413-aa72-4f0c-ab0b-f5c62d874948 for instance with vm_state building and task_state spawning. [ 1151.249925] env[62183]: DEBUG nova.network.neutron [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Successfully updated port: ae7b0413-aa72-4f0c-ab0b-f5c62d874948 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1151.752356] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.752503] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.752668] env[62183]: DEBUG nova.network.neutron [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1152.283875] env[62183]: DEBUG nova.network.neutron [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1152.414849] env[62183]: DEBUG nova.network.neutron [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Updating instance_info_cache with network_info: [{"id": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "address": "fa:16:3e:fa:b8:09", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae7b0413-aa", "ovs_interfaceid": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.917394] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.917705] env[62183]: DEBUG nova.compute.manager [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Instance network_info: |[{"id": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "address": "fa:16:3e:fa:b8:09", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae7b0413-aa", "ovs_interfaceid": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1152.918305] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:b8:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8800a981-a89e-42e4-8be9-cace419ba9cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae7b0413-aa72-4f0c-ab0b-f5c62d874948', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1152.925965] env[62183]: DEBUG oslo.service.loopingcall [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1152.926194] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1152.926817] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-488b8ea8-cf49-4e05-8b36-f07e0703bb62 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.947244] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1152.947244] env[62183]: value = "task-1387575" [ 1152.947244] env[62183]: _type = "Task" [ 1152.947244] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.954788] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387575, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.064522] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.064843] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.195854] env[62183]: DEBUG nova.compute.manager [req-60df1563-4572-40eb-bc65-832df51e5a6c req-4de1a54f-33eb-4af7-bc45-b92500dd68d1 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Received event network-changed-ae7b0413-aa72-4f0c-ab0b-f5c62d874948 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1153.196071] env[62183]: DEBUG nova.compute.manager [req-60df1563-4572-40eb-bc65-832df51e5a6c req-4de1a54f-33eb-4af7-bc45-b92500dd68d1 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Refreshing instance network info cache due to event network-changed-ae7b0413-aa72-4f0c-ab0b-f5c62d874948. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1153.196626] env[62183]: DEBUG oslo_concurrency.lockutils [req-60df1563-4572-40eb-bc65-832df51e5a6c req-4de1a54f-33eb-4af7-bc45-b92500dd68d1 service nova] Acquiring lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.196790] env[62183]: DEBUG oslo_concurrency.lockutils [req-60df1563-4572-40eb-bc65-832df51e5a6c req-4de1a54f-33eb-4af7-bc45-b92500dd68d1 service nova] Acquired lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.196961] env[62183]: DEBUG nova.network.neutron [req-60df1563-4572-40eb-bc65-832df51e5a6c req-4de1a54f-33eb-4af7-bc45-b92500dd68d1 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Refreshing network info cache for port ae7b0413-aa72-4f0c-ab0b-f5c62d874948 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1153.458405] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387575, 'name': CreateVM_Task, 'duration_secs': 0.289663} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.458807] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1153.459204] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.459384] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.459725] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1153.459977] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-388f59e4-9653-426b-95c4-2d09687d09ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.464356] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1153.464356] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52307d95-ca95-b02a-c969-016ed305c47e" [ 1153.464356] env[62183]: _type = "Task" [ 1153.464356] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.471427] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52307d95-ca95-b02a-c969-016ed305c47e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.567797] env[62183]: DEBUG nova.compute.utils [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1153.873275] env[62183]: DEBUG nova.network.neutron [req-60df1563-4572-40eb-bc65-832df51e5a6c req-4de1a54f-33eb-4af7-bc45-b92500dd68d1 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Updated VIF entry in instance network info cache for port ae7b0413-aa72-4f0c-ab0b-f5c62d874948. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1153.873632] env[62183]: DEBUG nova.network.neutron [req-60df1563-4572-40eb-bc65-832df51e5a6c req-4de1a54f-33eb-4af7-bc45-b92500dd68d1 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Updating instance_info_cache with network_info: [{"id": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "address": "fa:16:3e:fa:b8:09", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae7b0413-aa", "ovs_interfaceid": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.974789] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52307d95-ca95-b02a-c969-016ed305c47e, 'name': SearchDatastore_Task, 'duration_secs': 0.010547} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.975106] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.975372] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1153.975584] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.975733] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.975945] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1153.976171] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b0d74c0-a62b-482d-93f7-287bd087ad42 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.983703] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1153.983870] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1153.984544] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b4ee110-15ce-4908-90b7-360a440538a4 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.989146] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1153.989146] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52659420-0580-121d-d748-fb418d8bcbf1" [ 1153.989146] env[62183]: _type = "Task" [ 1153.989146] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.995788] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52659420-0580-121d-d748-fb418d8bcbf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.070892] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.376764] env[62183]: DEBUG oslo_concurrency.lockutils [req-60df1563-4572-40eb-bc65-832df51e5a6c req-4de1a54f-33eb-4af7-bc45-b92500dd68d1 service nova] Releasing lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.499270] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52659420-0580-121d-d748-fb418d8bcbf1, 'name': SearchDatastore_Task, 'duration_secs': 0.00756} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.500043] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37c739f3-0ee4-4c08-8e4e-1420b6a32aa0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.504775] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1154.504775] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52bb0e04-bb4c-df99-b942-0094feb5f99e" [ 1154.504775] env[62183]: _type = "Task" [ 1154.504775] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.512072] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bb0e04-bb4c-df99-b942-0094feb5f99e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.014677] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52bb0e04-bb4c-df99-b942-0094feb5f99e, 'name': SearchDatastore_Task, 'duration_secs': 0.008557} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.014987] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.015334] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 87961eb2-d977-48c1-9eaf-c371eac9bbb1/87961eb2-d977-48c1-9eaf-c371eac9bbb1.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1155.015612] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-870481b9-0a3d-4552-a688-c6e254401b18 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.023610] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1155.023610] env[62183]: value = "task-1387576" [ 1155.023610] env[62183]: _type = "Task" [ 1155.023610] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.030673] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.127747] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.128048] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.128305] env[62183]: INFO nova.compute.manager [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Attaching volume 0f1588a0-29c0-4210-8639-cbccdc187a46 to /dev/sdb [ 1155.158103] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8917a37b-8a6b-459b-890e-787389b0fd75 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.165510] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ff4cac-4f2a-457e-b5af-b3946c9ac10c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.179115] env[62183]: DEBUG nova.virt.block_device [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Updating existing volume attachment record: 42fe7c81-8540-4054-9127-8aea66258928 {{(pid=62183) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1155.533329] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387576, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480209} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.533687] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 87961eb2-d977-48c1-9eaf-c371eac9bbb1/87961eb2-d977-48c1-9eaf-c371eac9bbb1.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1155.533799] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1155.534072] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-efd81bcd-0836-4acd-a191-8e659bfe5135 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.540517] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1155.540517] env[62183]: value = "task-1387578" [ 1155.540517] env[62183]: _type = "Task" [ 1155.540517] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.548362] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387578, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.049743] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387578, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073915} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.050033] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1156.050821] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ffb5ce-2cef-4f27-ba13-e577b9e09f05 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.071949] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 87961eb2-d977-48c1-9eaf-c371eac9bbb1/87961eb2-d977-48c1-9eaf-c371eac9bbb1.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.072219] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-097d1be5-ffe9-444d-a308-711445f8146d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.091516] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1156.091516] env[62183]: value = "task-1387579" [ 1156.091516] env[62183]: _type = "Task" [ 1156.091516] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.098945] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387579, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.600802] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387579, 'name': ReconfigVM_Task, 'duration_secs': 0.342339} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.601151] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 87961eb2-d977-48c1-9eaf-c371eac9bbb1/87961eb2-d977-48c1-9eaf-c371eac9bbb1.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1156.601718] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b930df71-8fc0-4ded-b63b-c0d459fabf0d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.608414] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1156.608414] env[62183]: value = "task-1387580" [ 1156.608414] env[62183]: _type = "Task" [ 1156.608414] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.615507] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387580, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.118223] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387580, 'name': Rename_Task, 'duration_secs': 0.132759} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.118499] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1157.118774] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5fdba40-6532-45c4-a94f-6cf5e8152d9a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.124804] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1157.124804] env[62183]: value = "task-1387581" [ 1157.124804] env[62183]: _type = "Task" [ 1157.124804] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.134296] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387581, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.634380] env[62183]: DEBUG oslo_vmware.api [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387581, 'name': PowerOnVM_Task, 'duration_secs': 0.431247} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.634759] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1157.634974] env[62183]: INFO nova.compute.manager [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Took 6.62 seconds to spawn the instance on the hypervisor. [ 1157.635184] env[62183]: DEBUG nova.compute.manager [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1157.635966] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb34cec-6b1a-4792-a8a5-16c02320a5f1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.155223] env[62183]: INFO nova.compute.manager [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Took 14.43 seconds to build instance. [ 1158.657043] env[62183]: DEBUG oslo_concurrency.lockutils [None req-53c7788f-bb71-40ae-a2a6-7a719c38dfa7 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.936s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.971148] env[62183]: DEBUG nova.compute.manager [req-35308695-ca5c-4f5a-b5c9-b25083914a3a req-ba8757b6-09c5-4603-92ff-f285061b7bdd service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Received event network-changed-ae7b0413-aa72-4f0c-ab0b-f5c62d874948 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1158.971361] env[62183]: DEBUG nova.compute.manager [req-35308695-ca5c-4f5a-b5c9-b25083914a3a req-ba8757b6-09c5-4603-92ff-f285061b7bdd service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Refreshing instance network info cache due to event network-changed-ae7b0413-aa72-4f0c-ab0b-f5c62d874948. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1158.971629] env[62183]: DEBUG oslo_concurrency.lockutils [req-35308695-ca5c-4f5a-b5c9-b25083914a3a req-ba8757b6-09c5-4603-92ff-f285061b7bdd service nova] Acquiring lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.971734] env[62183]: DEBUG oslo_concurrency.lockutils [req-35308695-ca5c-4f5a-b5c9-b25083914a3a req-ba8757b6-09c5-4603-92ff-f285061b7bdd service nova] Acquired lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.971876] env[62183]: DEBUG nova.network.neutron [req-35308695-ca5c-4f5a-b5c9-b25083914a3a req-ba8757b6-09c5-4603-92ff-f285061b7bdd service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Refreshing network info cache for port ae7b0413-aa72-4f0c-ab0b-f5c62d874948 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1159.694438] env[62183]: DEBUG nova.network.neutron [req-35308695-ca5c-4f5a-b5c9-b25083914a3a req-ba8757b6-09c5-4603-92ff-f285061b7bdd service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Updated VIF entry in instance network info cache for port ae7b0413-aa72-4f0c-ab0b-f5c62d874948. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1159.695027] env[62183]: DEBUG nova.network.neutron [req-35308695-ca5c-4f5a-b5c9-b25083914a3a req-ba8757b6-09c5-4603-92ff-f285061b7bdd service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Updating instance_info_cache with network_info: [{"id": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "address": "fa:16:3e:fa:b8:09", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae7b0413-aa", "ovs_interfaceid": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.725994] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Volume attach. Driver type: vmdk {{(pid=62183) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1159.726239] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294530', 'volume_id': '0f1588a0-29c0-4210-8639-cbccdc187a46', 'name': 'volume-0f1588a0-29c0-4210-8639-cbccdc187a46', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '632ccdc6-d5fa-4e42-8273-c2e46eb1bf11', 'attached_at': '', 'detached_at': '', 'volume_id': '0f1588a0-29c0-4210-8639-cbccdc187a46', 'serial': '0f1588a0-29c0-4210-8639-cbccdc187a46'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1159.727126] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6d5a1a-7f89-4606-803f-eb3e5719625f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.744301] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971a9056-3a70-4cf5-b29b-a7a2631515d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.768012] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-0f1588a0-29c0-4210-8639-cbccdc187a46/volume-0f1588a0-29c0-4210-8639-cbccdc187a46.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.768273] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61836419-656d-4908-bab4-e433174d4be8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.785594] env[62183]: DEBUG oslo_vmware.api [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1159.785594] env[62183]: value = "task-1387583" [ 1159.785594] env[62183]: _type = "Task" [ 1159.785594] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.793214] env[62183]: DEBUG oslo_vmware.api [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387583, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.198201] env[62183]: DEBUG oslo_concurrency.lockutils [req-35308695-ca5c-4f5a-b5c9-b25083914a3a req-ba8757b6-09c5-4603-92ff-f285061b7bdd service nova] Releasing lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.296820] env[62183]: DEBUG oslo_vmware.api [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387583, 'name': ReconfigVM_Task, 'duration_secs': 0.305425} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.297069] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-0f1588a0-29c0-4210-8639-cbccdc187a46/volume-0f1588a0-29c0-4210-8639-cbccdc187a46.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.301732] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9a18761-9684-44f2-9826-acc5332833ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.316217] env[62183]: DEBUG oslo_vmware.api [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1160.316217] env[62183]: value = "task-1387584" [ 1160.316217] env[62183]: _type = "Task" [ 1160.316217] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.323686] env[62183]: DEBUG oslo_vmware.api [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387584, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.825889] env[62183]: DEBUG oslo_vmware.api [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387584, 'name': ReconfigVM_Task, 'duration_secs': 0.178463} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.826231] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294530', 'volume_id': '0f1588a0-29c0-4210-8639-cbccdc187a46', 'name': 'volume-0f1588a0-29c0-4210-8639-cbccdc187a46', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '632ccdc6-d5fa-4e42-8273-c2e46eb1bf11', 'attached_at': '', 'detached_at': '', 'volume_id': '0f1588a0-29c0-4210-8639-cbccdc187a46', 'serial': '0f1588a0-29c0-4210-8639-cbccdc187a46'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1161.862352] env[62183]: DEBUG nova.objects.instance [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'flavor' on Instance uuid 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.367361] env[62183]: DEBUG oslo_concurrency.lockutils [None req-9e62722a-70dc-48e1-9fe2-d81eb40bed94 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.239s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.541077] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.541386] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.044611] env[62183]: INFO nova.compute.manager [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Detaching volume 0f1588a0-29c0-4210-8639-cbccdc187a46 [ 1163.085950] env[62183]: INFO nova.virt.block_device [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Attempting to driver detach volume 0f1588a0-29c0-4210-8639-cbccdc187a46 from mountpoint /dev/sdb [ 1163.086401] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Volume detach. Driver type: vmdk {{(pid=62183) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1163.086725] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294530', 'volume_id': '0f1588a0-29c0-4210-8639-cbccdc187a46', 'name': 'volume-0f1588a0-29c0-4210-8639-cbccdc187a46', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '632ccdc6-d5fa-4e42-8273-c2e46eb1bf11', 'attached_at': '', 'detached_at': '', 'volume_id': '0f1588a0-29c0-4210-8639-cbccdc187a46', 'serial': '0f1588a0-29c0-4210-8639-cbccdc187a46'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1163.087971] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7014661-666b-467c-8e4e-9513c9372858 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.115132] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82bca03-70bc-4561-a93b-cfd9a8ac4a6b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.122197] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a716e0f-a644-4288-a808-bc71ab1f3d64 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.141565] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2efa74a-91c2-496f-b07c-940f81a8f86c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.156975] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] The volume has not been displaced from its original location: [datastore2] volume-0f1588a0-29c0-4210-8639-cbccdc187a46/volume-0f1588a0-29c0-4210-8639-cbccdc187a46.vmdk. No consolidation needed. {{(pid=62183) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1163.162251] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1163.162612] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f79fd4c4-5efa-4c5a-8017-2a8645e30b6f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.182149] env[62183]: DEBUG oslo_vmware.api [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1163.182149] env[62183]: value = "task-1387585" [ 1163.182149] env[62183]: _type = "Task" [ 1163.182149] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.189589] env[62183]: DEBUG oslo_vmware.api [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387585, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.691769] env[62183]: DEBUG oslo_vmware.api [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387585, 'name': ReconfigVM_Task, 'duration_secs': 0.194074} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.692097] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1163.696741] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3b58962-8815-4a0c-ba22-b193f3ff8413 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.711490] env[62183]: DEBUG oslo_vmware.api [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1163.711490] env[62183]: value = "task-1387586" [ 1163.711490] env[62183]: _type = "Task" [ 1163.711490] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.718956] env[62183]: DEBUG oslo_vmware.api [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387586, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.221540] env[62183]: DEBUG oslo_vmware.api [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387586, 'name': ReconfigVM_Task, 'duration_secs': 0.13392} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.221906] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294530', 'volume_id': '0f1588a0-29c0-4210-8639-cbccdc187a46', 'name': 'volume-0f1588a0-29c0-4210-8639-cbccdc187a46', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '632ccdc6-d5fa-4e42-8273-c2e46eb1bf11', 'attached_at': '', 'detached_at': '', 'volume_id': '0f1588a0-29c0-4210-8639-cbccdc187a46', 'serial': '0f1588a0-29c0-4210-8639-cbccdc187a46'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1164.763931] env[62183]: DEBUG nova.objects.instance [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'flavor' on Instance uuid 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.894700] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.894947] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.397786] env[62183]: DEBUG nova.compute.utils [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1165.772587] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c4d3aad9-b021-4218-a201-73555ca57c71 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.231s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.900468] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.822636] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.823035] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.823155] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.823358] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.823540] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.826135] env[62183]: INFO nova.compute.manager [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Terminating instance [ 1166.827825] env[62183]: DEBUG nova.compute.manager [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1166.828038] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1166.828863] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ec2196-f6a2-4cdc-a391-bffae2709492 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.836234] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1166.836456] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23ef9b8d-858d-4a9c-bcaf-fd901a0b8735 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.842718] env[62183]: DEBUG oslo_vmware.api [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1166.842718] env[62183]: value = "task-1387587" [ 1166.842718] env[62183]: _type = "Task" [ 1166.842718] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.850739] env[62183]: DEBUG oslo_vmware.api [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.975862] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.976149] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.976409] env[62183]: INFO nova.compute.manager [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Attaching volume f1bf2e91-7d66-4799-9d7b-ed6869cbc01c to /dev/sdb [ 1167.007278] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0a4c84-4769-4501-90f0-c3ba7ac75f70 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.014284] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116e11d1-a316-4cb3-a7a9-110f902537ca {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.027079] env[62183]: DEBUG nova.virt.block_device [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Updating existing volume attachment record: 5f13c006-a8b7-4de8-b2ba-ffc1e39789b0 {{(pid=62183) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1167.352552] env[62183]: DEBUG oslo_vmware.api [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387587, 'name': PowerOffVM_Task, 'duration_secs': 0.185016} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.352827] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1167.352998] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1167.353271] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4f91ad2-0aff-4805-b161-a7c5600ff3d0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.414679] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1167.414952] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1167.415158] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleting the datastore file [datastore1] 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1167.415418] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d229fea6-035d-4701-9f51-b7a862f4b516 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.421670] env[62183]: DEBUG oslo_vmware.api [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for the task: (returnval){ [ 1167.421670] env[62183]: value = "task-1387590" [ 1167.421670] env[62183]: _type = "Task" [ 1167.421670] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.428913] env[62183]: DEBUG oslo_vmware.api [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387590, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.931703] env[62183]: DEBUG oslo_vmware.api [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Task: {'id': task-1387590, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205198} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.932074] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1167.932128] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1167.932291] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1167.932468] env[62183]: INFO nova.compute.manager [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1167.932711] env[62183]: DEBUG oslo.service.loopingcall [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1167.932910] env[62183]: DEBUG nova.compute.manager [-] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1167.933016] env[62183]: DEBUG nova.network.neutron [-] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1168.348918] env[62183]: DEBUG nova.compute.manager [req-16f6a614-07ce-4705-b38c-46060d4d7eb7 req-5a9bfedd-56a5-4bf3-a8ad-6236714a7d22 service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Received event network-vif-deleted-c7b92447-a61f-4d71-b39f-28af011a2142 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1168.349146] env[62183]: INFO nova.compute.manager [req-16f6a614-07ce-4705-b38c-46060d4d7eb7 req-5a9bfedd-56a5-4bf3-a8ad-6236714a7d22 service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Neutron deleted interface c7b92447-a61f-4d71-b39f-28af011a2142; detaching it from the instance and deleting it from the info cache [ 1168.349327] env[62183]: DEBUG nova.network.neutron [req-16f6a614-07ce-4705-b38c-46060d4d7eb7 req-5a9bfedd-56a5-4bf3-a8ad-6236714a7d22 service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.820279] env[62183]: DEBUG nova.network.neutron [-] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.852289] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aebd8b69-969d-481b-95d6-1119f86aa534 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.861248] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c70298-7f87-4667-955d-6bb34e31f4a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.887826] env[62183]: DEBUG nova.compute.manager [req-16f6a614-07ce-4705-b38c-46060d4d7eb7 req-5a9bfedd-56a5-4bf3-a8ad-6236714a7d22 service nova] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Detach interface failed, port_id=c7b92447-a61f-4d71-b39f-28af011a2142, reason: Instance 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1169.323631] env[62183]: INFO nova.compute.manager [-] [instance: 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11] Took 1.39 seconds to deallocate network for instance. [ 1169.830493] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.830798] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.831076] env[62183]: DEBUG nova.objects.instance [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lazy-loading 'resources' on Instance uuid 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.391976] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3152f550-743b-4d3b-b46b-f16418f25f59 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.399207] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc0d8e4-4160-42ac-a1c8-bc6179354646 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.428942] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b73f58-bd1e-46e5-b78e-d9df0754ccd1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.435781] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de07ec3-f63f-43f5-85b2-a63606de296e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.448367] env[62183]: DEBUG nova.compute.provider_tree [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1170.977058] env[62183]: DEBUG nova.scheduler.client.report [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 147 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1170.977330] env[62183]: DEBUG nova.compute.provider_tree [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 147 to 148 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1170.977512] env[62183]: DEBUG nova.compute.provider_tree [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1171.445018] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.445362] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.445553] env[62183]: INFO nova.compute.manager [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Shelving [ 1171.482407] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.652s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.500658] env[62183]: INFO nova.scheduler.client.report [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Deleted allocations for instance 632ccdc6-d5fa-4e42-8273-c2e46eb1bf11 [ 1171.571258] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Volume attach. Driver type: vmdk {{(pid=62183) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1171.571514] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294531', 'volume_id': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'name': 'volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1b411fab-7728-4b29-ad9f-ce22a2face55', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'serial': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1171.572423] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3d8dab-b40d-4db1-869b-154e1e991727 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.588998] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76164215-1e9f-498e-a8f4-8ef6788aa26e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.613123] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c/volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1171.613418] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc7e3a6f-aa0a-4562-a462-9cada9abff37 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.631569] env[62183]: DEBUG oslo_vmware.api [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1171.631569] env[62183]: value = "task-1387592" [ 1171.631569] env[62183]: _type = "Task" [ 1171.631569] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.638736] env[62183]: DEBUG oslo_vmware.api [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387592, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.952849] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1171.953203] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f431fe1d-4d3b-4e1c-be3e-8a14141335ef {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.960101] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1171.960101] env[62183]: value = "task-1387593" [ 1171.960101] env[62183]: _type = "Task" [ 1171.960101] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.968229] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387593, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.008333] env[62183]: DEBUG oslo_concurrency.lockutils [None req-b54943db-3bdc-4d9c-b2a6-20cc99fe6135 tempest-AttachVolumeNegativeTest-691436208 tempest-AttachVolumeNegativeTest-691436208-project-member] Lock "632ccdc6-d5fa-4e42-8273-c2e46eb1bf11" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.185s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.140819] env[62183]: DEBUG oslo_vmware.api [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387592, 'name': ReconfigVM_Task, 'duration_secs': 0.326195} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.141108] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c/volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1172.145938] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3b53a17-fe96-4ee5-9d5f-4c9133ddb4ac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.160325] env[62183]: DEBUG oslo_vmware.api [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1172.160325] env[62183]: value = "task-1387594" [ 1172.160325] env[62183]: _type = "Task" [ 1172.160325] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.168238] env[62183]: DEBUG oslo_vmware.api [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.469457] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387593, 'name': PowerOffVM_Task, 'duration_secs': 0.197853} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.469892] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1172.470469] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cebeee-4ce9-4c50-b329-41a60b6e1d6a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.489264] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7338ad25-5c59-4e46-af5f-ac4354505ca1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.669606] env[62183]: DEBUG oslo_vmware.api [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387594, 'name': ReconfigVM_Task, 'duration_secs': 0.131118} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.669942] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294531', 'volume_id': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'name': 'volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1b411fab-7728-4b29-ad9f-ce22a2face55', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'serial': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1172.998807] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Creating Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1172.999159] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-56ac786b-4110-438f-90ed-f74fb9ba663c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.007130] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1173.007130] env[62183]: value = "task-1387596" [ 1173.007130] env[62183]: _type = "Task" [ 1173.007130] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.015308] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387596, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.518904] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387596, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.703092] env[62183]: DEBUG nova.objects.instance [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 1b411fab-7728-4b29-ad9f-ce22a2face55 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1174.017572] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387596, 'name': CreateSnapshot_Task, 'duration_secs': 0.561847} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.017867] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Created Snapshot of the VM instance {{(pid=62183) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1174.018614] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc24cac-5003-4782-859b-6422cfb6ec12 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.207853] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ed3d4de6-151a-44bf-ba71-4db08adf2b9b tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.232s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.534763] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Creating linked-clone VM from snapshot {{(pid=62183) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1174.536034] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-569544a9-3954-4a4e-aa4f-1d1921b8fe11 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.544880] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1174.544880] env[62183]: value = "task-1387597" [ 1174.544880] env[62183]: _type = "Task" [ 1174.544880] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.552601] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387597, 'name': CloneVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.038292] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.038536] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1175.055644] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387597, 'name': CloneVM_Task} progress is 94%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.541887] env[62183]: DEBUG nova.compute.utils [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1175.557037] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387597, 'name': CloneVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.044933] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.055062] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387597, 'name': CloneVM_Task, 'duration_secs': 1.059913} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.055314] env[62183]: INFO nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Created linked-clone VM from snapshot [ 1176.056018] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b7c002-8086-4773-91f6-f2d43ba9ef4a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.062809] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Uploading image 72a931cc-90a1-49b8-8a97-5970cd9b0d17 {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1176.087502] env[62183]: DEBUG oslo_vmware.rw_handles [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1176.087502] env[62183]: value = "vm-294533" [ 1176.087502] env[62183]: _type = "VirtualMachine" [ 1176.087502] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1176.087813] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b8b71753-fc3d-4c6f-84b0-497effb1f7d9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.095316] env[62183]: DEBUG oslo_vmware.rw_handles [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease: (returnval){ [ 1176.095316] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52eaa3ea-5371-59d8-d759-4c2336087d47" [ 1176.095316] env[62183]: _type = "HttpNfcLease" [ 1176.095316] env[62183]: } obtained for exporting VM: (result){ [ 1176.095316] env[62183]: value = "vm-294533" [ 1176.095316] env[62183]: _type = "VirtualMachine" [ 1176.095316] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1176.095607] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the lease: (returnval){ [ 1176.095607] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52eaa3ea-5371-59d8-d759-4c2336087d47" [ 1176.095607] env[62183]: _type = "HttpNfcLease" [ 1176.095607] env[62183]: } to be ready. {{(pid=62183) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1176.101382] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1176.101382] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52eaa3ea-5371-59d8-d759-4c2336087d47" [ 1176.101382] env[62183]: _type = "HttpNfcLease" [ 1176.101382] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1176.608842] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1176.608842] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52eaa3ea-5371-59d8-d759-4c2336087d47" [ 1176.608842] env[62183]: _type = "HttpNfcLease" [ 1176.608842] env[62183]: } is ready. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1176.609429] env[62183]: DEBUG oslo_vmware.rw_handles [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1176.609429] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52eaa3ea-5371-59d8-d759-4c2336087d47" [ 1176.609429] env[62183]: _type = "HttpNfcLease" [ 1176.609429] env[62183]: }. {{(pid=62183) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1176.609965] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9610f8c6-9a34-4dc7-bb46-65cd52f2bd94 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.618914] env[62183]: DEBUG oslo_vmware.rw_handles [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a6500-880f-ecd3-1ef4-13c87c49c165/disk-0.vmdk from lease info. {{(pid=62183) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1176.619131] env[62183]: DEBUG oslo_vmware.rw_handles [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a6500-880f-ecd3-1ef4-13c87c49c165/disk-0.vmdk for reading. {{(pid=62183) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1176.706481] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8156c3a5-ae35-460a-ad79-65445b874cc9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.105913] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.106440] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.106777] env[62183]: INFO nova.compute.manager [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Attaching volume 500aebf0-043a-4439-b44e-60a6d2001622 to /dev/sdc [ 1177.139761] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dbc0940-a5fb-4fca-af78-50dd48de6bff {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.147386] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916a42cc-4b26-43e7-b475-e07e35e7f7f8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.160999] env[62183]: DEBUG nova.virt.block_device [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Updating existing volume attachment record: cc7ae603-7ab1-4ec4-b3ce-5acd06d472c0 {{(pid=62183) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1181.705036] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Volume attach. Driver type: vmdk {{(pid=62183) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1181.705411] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294534', 'volume_id': '500aebf0-043a-4439-b44e-60a6d2001622', 'name': 'volume-500aebf0-043a-4439-b44e-60a6d2001622', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1b411fab-7728-4b29-ad9f-ce22a2face55', 'attached_at': '', 'detached_at': '', 'volume_id': '500aebf0-043a-4439-b44e-60a6d2001622', 'serial': '500aebf0-043a-4439-b44e-60a6d2001622'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1181.706458] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dbd29cf-0abf-47f3-88a1-4fddfcb52024 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.723097] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0961db0a-e5cc-47a5-9b21-3314aaa9e2cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.753654] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] volume-500aebf0-043a-4439-b44e-60a6d2001622/volume-500aebf0-043a-4439-b44e-60a6d2001622.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1181.753992] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3c4a0f0-f227-4fa5-95ce-d78f48e4fc8e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.772729] env[62183]: DEBUG oslo_vmware.api [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1181.772729] env[62183]: value = "task-1387603" [ 1181.772729] env[62183]: _type = "Task" [ 1181.772729] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.781180] env[62183]: DEBUG oslo_vmware.api [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387603, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.282841] env[62183]: DEBUG oslo_vmware.api [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387603, 'name': ReconfigVM_Task, 'duration_secs': 0.377882} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.283158] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfigured VM instance instance-0000006c to attach disk [datastore2] volume-500aebf0-043a-4439-b44e-60a6d2001622/volume-500aebf0-043a-4439-b44e-60a6d2001622.vmdk or device None with type thin {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1182.287868] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4c4bcf7-3fd6-46ab-ba6a-b32b9dfd5e0c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.302232] env[62183]: DEBUG oslo_vmware.api [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1182.302232] env[62183]: value = "task-1387604" [ 1182.302232] env[62183]: _type = "Task" [ 1182.302232] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.310108] env[62183]: DEBUG oslo_vmware.api [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387604, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.813474] env[62183]: DEBUG oslo_vmware.api [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387604, 'name': ReconfigVM_Task, 'duration_secs': 0.133604} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.813852] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294534', 'volume_id': '500aebf0-043a-4439-b44e-60a6d2001622', 'name': 'volume-500aebf0-043a-4439-b44e-60a6d2001622', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1b411fab-7728-4b29-ad9f-ce22a2face55', 'attached_at': '', 'detached_at': '', 'volume_id': '500aebf0-043a-4439-b44e-60a6d2001622', 'serial': '500aebf0-043a-4439-b44e-60a6d2001622'} {{(pid=62183) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1183.860689] env[62183]: DEBUG oslo_vmware.rw_handles [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a6500-880f-ecd3-1ef4-13c87c49c165/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1183.861695] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94250f50-e092-4a98-b8f7-fdfa99524658 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.864929] env[62183]: DEBUG nova.objects.instance [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 1b411fab-7728-4b29-ad9f-ce22a2face55 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.871013] env[62183]: DEBUG oslo_vmware.rw_handles [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a6500-880f-ecd3-1ef4-13c87c49c165/disk-0.vmdk is in state: ready. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1183.871013] env[62183]: ERROR oslo_vmware.rw_handles [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a6500-880f-ecd3-1ef4-13c87c49c165/disk-0.vmdk due to incomplete transfer. [ 1183.871013] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-26d65ab7-4677-45c1-a5d0-fca0d82fdc30 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.880018] env[62183]: DEBUG oslo_vmware.rw_handles [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a6500-880f-ecd3-1ef4-13c87c49c165/disk-0.vmdk. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1183.880018] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Uploaded image 72a931cc-90a1-49b8-8a97-5970cd9b0d17 to the Glance image server {{(pid=62183) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1183.881836] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Destroying the VM {{(pid=62183) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1183.882705] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6aef8119-cd4f-4a1c-865b-acb8aaffa084 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.888129] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1183.888129] env[62183]: value = "task-1387605" [ 1183.888129] env[62183]: _type = "Task" [ 1183.888129] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.896123] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387605, 'name': Destroy_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.373748] env[62183]: DEBUG oslo_concurrency.lockutils [None req-07ecd14c-a45f-4bff-974a-64f79849a5a3 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.267s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.398009] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387605, 'name': Destroy_Task, 'duration_secs': 0.315598} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.398274] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Destroyed the VM [ 1184.398523] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Deleting Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1184.398767] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a40c8145-294f-4452-b17e-5e1ce7473552 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.404928] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1184.404928] env[62183]: value = "task-1387606" [ 1184.404928] env[62183]: _type = "Task" [ 1184.404928] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.412534] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387606, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.672822] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.673167] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.915064] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387606, 'name': RemoveSnapshot_Task, 'duration_secs': 0.340851} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.915443] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Deleted Snapshot of the VM instance {{(pid=62183) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1184.915640] env[62183]: DEBUG nova.compute.manager [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.916467] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2620ac8-8507-41f4-9761-66215d8d88a9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.176822] env[62183]: INFO nova.compute.manager [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Detaching volume f1bf2e91-7d66-4799-9d7b-ed6869cbc01c [ 1185.212960] env[62183]: INFO nova.virt.block_device [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Attempting to driver detach volume f1bf2e91-7d66-4799-9d7b-ed6869cbc01c from mountpoint /dev/sdb [ 1185.213232] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Volume detach. Driver type: vmdk {{(pid=62183) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1185.213424] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294531', 'volume_id': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'name': 'volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1b411fab-7728-4b29-ad9f-ce22a2face55', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'serial': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1185.214350] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6060274-905e-46b3-8714-b49fdb487899 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.239426] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4b7e3b-6fb8-4efe-99d8-a4ead11654c1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.247012] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a04347-9a1b-4f66-b849-095fe41bae2d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.268762] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c481765-1f6d-4247-ad41-dc85ad11999c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.282638] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] The volume has not been displaced from its original location: [datastore2] volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c/volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c.vmdk. No consolidation needed. {{(pid=62183) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1185.287745] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1185.288059] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ba3de0e-fe98-41b6-9cdc-e19662f6251d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.305685] env[62183]: DEBUG oslo_vmware.api [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1185.305685] env[62183]: value = "task-1387607" [ 1185.305685] env[62183]: _type = "Task" [ 1185.305685] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.317831] env[62183]: DEBUG oslo_vmware.api [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387607, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.428321] env[62183]: INFO nova.compute.manager [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Shelve offloading [ 1185.430010] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.430293] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f1c0480-1898-41c9-91de-55136ac5d6ac {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.437182] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1185.437182] env[62183]: value = "task-1387608" [ 1185.437182] env[62183]: _type = "Task" [ 1185.437182] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.444809] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387608, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.815637] env[62183]: DEBUG oslo_vmware.api [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387607, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.948060] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] VM already powered off {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1185.948449] env[62183]: DEBUG nova.compute.manager [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1185.949075] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8ddf5b-2bd7-4c22-90c6-657a3c7eb447 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.954724] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.954891] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.955079] env[62183]: DEBUG nova.network.neutron [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1186.316090] env[62183]: DEBUG oslo_vmware.api [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387607, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.658938] env[62183]: DEBUG nova.network.neutron [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating instance_info_cache with network_info: [{"id": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "address": "fa:16:3e:80:88:09", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ebdb1ac-70", "ovs_interfaceid": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.818051] env[62183]: DEBUG oslo_vmware.api [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387607, 'name': ReconfigVM_Task, 'duration_secs': 1.234514} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.818051] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1186.822042] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e61c3a15-998a-45a9-ad59-4f2baa636068 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.836553] env[62183]: DEBUG oslo_vmware.api [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1186.836553] env[62183]: value = "task-1387609" [ 1186.836553] env[62183]: _type = "Task" [ 1186.836553] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.844109] env[62183]: DEBUG oslo_vmware.api [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387609, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.161870] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1187.345909] env[62183]: DEBUG oslo_vmware.api [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387609, 'name': ReconfigVM_Task, 'duration_secs': 0.136213} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.346224] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294531', 'volume_id': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'name': 'volume-f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1b411fab-7728-4b29-ad9f-ce22a2face55', 'attached_at': '', 'detached_at': '', 'volume_id': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c', 'serial': 'f1bf2e91-7d66-4799-9d7b-ed6869cbc01c'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1187.370524] env[62183]: DEBUG nova.compute.manager [req-42b749d9-a6c4-4287-b12e-0f3a946e333e req-2aedd0ea-f50f-4ca1-b8dc-d030e1f33c3a service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received event network-vif-unplugged-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1187.370750] env[62183]: DEBUG oslo_concurrency.lockutils [req-42b749d9-a6c4-4287-b12e-0f3a946e333e req-2aedd0ea-f50f-4ca1-b8dc-d030e1f33c3a service nova] Acquiring lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1187.370960] env[62183]: DEBUG oslo_concurrency.lockutils [req-42b749d9-a6c4-4287-b12e-0f3a946e333e req-2aedd0ea-f50f-4ca1-b8dc-d030e1f33c3a service nova] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.371146] env[62183]: DEBUG oslo_concurrency.lockutils [req-42b749d9-a6c4-4287-b12e-0f3a946e333e req-2aedd0ea-f50f-4ca1-b8dc-d030e1f33c3a service nova] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.371316] env[62183]: DEBUG nova.compute.manager [req-42b749d9-a6c4-4287-b12e-0f3a946e333e req-2aedd0ea-f50f-4ca1-b8dc-d030e1f33c3a service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] No waiting events found dispatching network-vif-unplugged-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1187.371491] env[62183]: WARNING nova.compute.manager [req-42b749d9-a6c4-4287-b12e-0f3a946e333e req-2aedd0ea-f50f-4ca1-b8dc-d030e1f33c3a service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received unexpected event network-vif-unplugged-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a for instance with vm_state shelved and task_state shelving_offloading. [ 1187.464544] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1187.465457] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b35a63-7c66-4591-a6ef-522480802266 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.473170] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1187.473403] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7bda9dcf-5bcb-4955-a945-6b895a876d91 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.544646] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1187.544868] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1187.545075] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleting the datastore file [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.545343] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-046a82ad-87f4-430d-aca6-9a481cc5fad9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.551297] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1187.551297] env[62183]: value = "task-1387611" [ 1187.551297] env[62183]: _type = "Task" [ 1187.551297] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.559045] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387611, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.885514] env[62183]: DEBUG nova.objects.instance [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 1b411fab-7728-4b29-ad9f-ce22a2face55 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.060760] env[62183]: DEBUG oslo_vmware.api [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387611, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1207} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.060984] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1188.061168] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1188.061349] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1188.089603] env[62183]: INFO nova.scheduler.client.report [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted allocations for instance 2d6f015c-5189-4a0b-bb79-3180d557e8f9 [ 1188.593903] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.594259] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.594376] env[62183]: DEBUG nova.objects.instance [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'resources' on Instance uuid 2d6f015c-5189-4a0b-bb79-3180d557e8f9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.893277] env[62183]: DEBUG oslo_concurrency.lockutils [None req-fedc328a-cb93-4557-b122-235063cc80a6 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.220s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.941094] env[62183]: DEBUG oslo_concurrency.lockutils [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.941529] env[62183]: DEBUG oslo_concurrency.lockutils [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.096744] env[62183]: DEBUG nova.objects.instance [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'numa_topology' on Instance uuid 2d6f015c-5189-4a0b-bb79-3180d557e8f9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.400668] env[62183]: DEBUG nova.compute.manager [req-01962016-50cd-48b9-a864-2b48d0b73ae0 req-60a8e10f-1567-4600-b494-c83f0f18c290 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received event network-changed-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1189.400852] env[62183]: DEBUG nova.compute.manager [req-01962016-50cd-48b9-a864-2b48d0b73ae0 req-60a8e10f-1567-4600-b494-c83f0f18c290 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Refreshing instance network info cache due to event network-changed-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1189.401083] env[62183]: DEBUG oslo_concurrency.lockutils [req-01962016-50cd-48b9-a864-2b48d0b73ae0 req-60a8e10f-1567-4600-b494-c83f0f18c290 service nova] Acquiring lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.401239] env[62183]: DEBUG oslo_concurrency.lockutils [req-01962016-50cd-48b9-a864-2b48d0b73ae0 req-60a8e10f-1567-4600-b494-c83f0f18c290 service nova] Acquired lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.401431] env[62183]: DEBUG nova.network.neutron [req-01962016-50cd-48b9-a864-2b48d0b73ae0 req-60a8e10f-1567-4600-b494-c83f0f18c290 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Refreshing network info cache for port 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1189.445320] env[62183]: INFO nova.compute.manager [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Detaching volume 500aebf0-043a-4439-b44e-60a6d2001622 [ 1189.478680] env[62183]: INFO nova.virt.block_device [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Attempting to driver detach volume 500aebf0-043a-4439-b44e-60a6d2001622 from mountpoint /dev/sdc [ 1189.478958] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Volume detach. Driver type: vmdk {{(pid=62183) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1189.479187] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294534', 'volume_id': '500aebf0-043a-4439-b44e-60a6d2001622', 'name': 'volume-500aebf0-043a-4439-b44e-60a6d2001622', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1b411fab-7728-4b29-ad9f-ce22a2face55', 'attached_at': '', 'detached_at': '', 'volume_id': '500aebf0-043a-4439-b44e-60a6d2001622', 'serial': '500aebf0-043a-4439-b44e-60a6d2001622'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1189.480063] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5795ed63-3cda-4752-b474-0703b124ffa7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.501364] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1b581b-4bea-4d15-a6ad-fa1cee3ca083 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.508495] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d267579-c9d9-4014-89c0-b773f2558ae7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.529521] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df23aa60-9615-4fcc-86e6-5d0836045d0b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.543724] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] The volume has not been displaced from its original location: [datastore2] volume-500aebf0-043a-4439-b44e-60a6d2001622/volume-500aebf0-043a-4439-b44e-60a6d2001622.vmdk. No consolidation needed. {{(pid=62183) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1189.549043] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfiguring VM instance instance-0000006c to detach disk 2002 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1189.549311] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e59a984-93ae-498f-ba50-9a69c4de9f54 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.566841] env[62183]: DEBUG oslo_vmware.api [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1189.566841] env[62183]: value = "task-1387612" [ 1189.566841] env[62183]: _type = "Task" [ 1189.566841] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.577778] env[62183]: DEBUG oslo_vmware.api [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387612, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.599459] env[62183]: DEBUG nova.objects.base [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Object Instance<2d6f015c-5189-4a0b-bb79-3180d557e8f9> lazy-loaded attributes: resources,numa_topology {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1189.644352] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f500e729-697d-4ba6-9cb1-a238e94732fd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.651426] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a1f0e5-1ad3-413e-a46a-5df8d5b27d8c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.680221] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4eb640-d7c0-4144-952d-8773b7d1754f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.686618] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d985396b-3aa8-4e31-986d-e6245bb0c38c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.699168] env[62183]: DEBUG nova.compute.provider_tree [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.076059] env[62183]: DEBUG oslo_vmware.api [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387612, 'name': ReconfigVM_Task, 'duration_secs': 0.224071} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.076385] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Reconfigured VM instance instance-0000006c to detach disk 2002 {{(pid=62183) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1190.080889] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca1aebf9-fc44-4920-b65e-3109cf800925 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.096098] env[62183]: DEBUG oslo_vmware.api [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1190.096098] env[62183]: value = "task-1387613" [ 1190.096098] env[62183]: _type = "Task" [ 1190.096098] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.103347] env[62183]: DEBUG oslo_vmware.api [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387613, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.122695] env[62183]: DEBUG nova.network.neutron [req-01962016-50cd-48b9-a864-2b48d0b73ae0 req-60a8e10f-1567-4600-b494-c83f0f18c290 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updated VIF entry in instance network info cache for port 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1190.123041] env[62183]: DEBUG nova.network.neutron [req-01962016-50cd-48b9-a864-2b48d0b73ae0 req-60a8e10f-1567-4600-b494-c83f0f18c290 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating instance_info_cache with network_info: [{"id": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "address": "fa:16:3e:80:88:09", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4ebdb1ac-70", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.202110] env[62183]: DEBUG nova.scheduler.client.report [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1190.383220] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.607056] env[62183]: DEBUG oslo_vmware.api [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387613, 'name': ReconfigVM_Task, 'duration_secs': 0.239544} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.607056] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-294534', 'volume_id': '500aebf0-043a-4439-b44e-60a6d2001622', 'name': 'volume-500aebf0-043a-4439-b44e-60a6d2001622', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '1b411fab-7728-4b29-ad9f-ce22a2face55', 'attached_at': '', 'detached_at': '', 'volume_id': '500aebf0-043a-4439-b44e-60a6d2001622', 'serial': '500aebf0-043a-4439-b44e-60a6d2001622'} {{(pid=62183) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1190.625380] env[62183]: DEBUG oslo_concurrency.lockutils [req-01962016-50cd-48b9-a864-2b48d0b73ae0 req-60a8e10f-1567-4600-b494-c83f0f18c290 service nova] Releasing lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.706614] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.112s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.147338] env[62183]: DEBUG nova.objects.instance [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'flavor' on Instance uuid 1b411fab-7728-4b29-ad9f-ce22a2face55 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.214911] env[62183]: DEBUG oslo_concurrency.lockutils [None req-0ecbc1e6-5065-4fbb-9636-af5c519a94a5 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.769s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.215739] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.833s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.215952] env[62183]: INFO nova.compute.manager [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Unshelving [ 1192.154889] env[62183]: DEBUG oslo_concurrency.lockutils [None req-04b28a77-ff0c-400a-926f-c0911abde054 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.213s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.240861] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1192.241159] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1192.241377] env[62183]: DEBUG nova.objects.instance [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'pci_requests' on Instance uuid 2d6f015c-5189-4a0b-bb79-3180d557e8f9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.745938] env[62183]: DEBUG nova.objects.instance [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'numa_topology' on Instance uuid 2d6f015c-5189-4a0b-bb79-3180d557e8f9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.248336] env[62183]: INFO nova.compute.claims [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1193.320321] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.320608] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.320879] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "1b411fab-7728-4b29-ad9f-ce22a2face55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.321055] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.321212] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.323761] env[62183]: INFO nova.compute.manager [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Terminating instance [ 1193.325595] env[62183]: DEBUG nova.compute.manager [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1193.325969] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1193.327122] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15787b4b-d0ac-4887-bb14-d23a7a123ff2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.335168] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1193.335447] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a51f534-01ea-4371-bfce-65c8fb83f535 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.342137] env[62183]: DEBUG oslo_vmware.api [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1193.342137] env[62183]: value = "task-1387614" [ 1193.342137] env[62183]: _type = "Task" [ 1193.342137] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.350578] env[62183]: DEBUG oslo_vmware.api [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387614, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.851658] env[62183]: DEBUG oslo_vmware.api [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387614, 'name': PowerOffVM_Task, 'duration_secs': 0.176623} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.851923] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1193.852116] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1193.852375] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c4cabc9-a36e-4fb6-a2da-a2e76e4b7d5b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.917510] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1193.917757] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1193.917974] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Deleting the datastore file [datastore1] 1b411fab-7728-4b29-ad9f-ce22a2face55 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1193.918276] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c148929b-2f83-45a3-aa23-ff1267f9a634 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.924506] env[62183]: DEBUG oslo_vmware.api [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for the task: (returnval){ [ 1193.924506] env[62183]: value = "task-1387616" [ 1193.924506] env[62183]: _type = "Task" [ 1193.924506] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.932268] env[62183]: DEBUG oslo_vmware.api [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387616, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.304088] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8c4e73-1fe3-416d-bae8-9f2fedf8bb9b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.311752] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce1c191-f594-48de-ad4e-45ba1190ea87 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.340011] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d56df0-07ee-40d4-81a1-b46c57cb242d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.346462] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc6f8c0-6ad3-4204-9cf2-ab4222bd52a2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.361716] env[62183]: DEBUG nova.compute.provider_tree [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1194.433889] env[62183]: DEBUG oslo_vmware.api [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Task: {'id': task-1387616, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133647} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.434158] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1194.434352] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1194.434534] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1194.434711] env[62183]: INFO nova.compute.manager [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1194.434988] env[62183]: DEBUG oslo.service.loopingcall [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1194.435498] env[62183]: DEBUG nova.compute.manager [-] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1194.435574] env[62183]: DEBUG nova.network.neutron [-] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1194.882152] env[62183]: ERROR nova.scheduler.client.report [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [req-decf31b8-20d3-48a8-a988-f70d588db3a1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-decf31b8-20d3-48a8-a988-f70d588db3a1"}]} [ 1194.897945] env[62183]: DEBUG nova.scheduler.client.report [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1194.912476] env[62183]: DEBUG nova.scheduler.client.report [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1194.912712] env[62183]: DEBUG nova.compute.provider_tree [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1194.923654] env[62183]: DEBUG nova.scheduler.client.report [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1194.945143] env[62183]: DEBUG nova.scheduler.client.report [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1194.988620] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd719e01-e1f0-479f-9206-3a619c226510 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.995970] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8105ba52-66bd-4e81-95d4-9ec424812b1f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.026430] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d95917-3d8e-4c84-b49b-ef1900fd0d61 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.033679] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5d2196-15ea-4d0c-8ca3-818e2b69bbbc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.047789] env[62183]: DEBUG nova.compute.provider_tree [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1195.050762] env[62183]: DEBUG nova.compute.manager [req-f03f902a-617f-46a2-928a-4dc594ba3fab req-4fca39e2-3ee1-4a08-8e45-25e535676ceb service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Received event network-vif-deleted-2de4249d-027d-4a0c-aab3-bc1828b18825 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1195.050940] env[62183]: INFO nova.compute.manager [req-f03f902a-617f-46a2-928a-4dc594ba3fab req-4fca39e2-3ee1-4a08-8e45-25e535676ceb service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Neutron deleted interface 2de4249d-027d-4a0c-aab3-bc1828b18825; detaching it from the instance and deleting it from the info cache [ 1195.051120] env[62183]: DEBUG nova.network.neutron [req-f03f902a-617f-46a2-928a-4dc594ba3fab req-4fca39e2-3ee1-4a08-8e45-25e535676ceb service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.530440] env[62183]: DEBUG nova.network.neutron [-] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.555740] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1aad7c9-2bd1-4a84-9483-7d4fc45338e1 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.565742] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a798b4aa-f573-47ba-8967-e43261ddec75 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.579699] env[62183]: DEBUG nova.scheduler.client.report [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 149 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1195.579968] env[62183]: DEBUG nova.compute.provider_tree [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 149 to 150 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1195.580241] env[62183]: DEBUG nova.compute.provider_tree [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1195.589802] env[62183]: DEBUG nova.compute.manager [req-f03f902a-617f-46a2-928a-4dc594ba3fab req-4fca39e2-3ee1-4a08-8e45-25e535676ceb service nova] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Detach interface failed, port_id=2de4249d-027d-4a0c-aab3-bc1828b18825, reason: Instance 1b411fab-7728-4b29-ad9f-ce22a2face55 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1196.032702] env[62183]: INFO nova.compute.manager [-] [instance: 1b411fab-7728-4b29-ad9f-ce22a2face55] Took 1.60 seconds to deallocate network for instance. [ 1196.084825] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.844s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.113930] env[62183]: INFO nova.network.neutron [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating port 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1196.407214] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.407498] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.407684] env[62183]: DEBUG nova.compute.manager [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1196.408641] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8a2e3c-1ae5-4d53-8177-195c9a7872cf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.415606] env[62183]: DEBUG nova.compute.manager [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62183) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1196.416167] env[62183]: DEBUG nova.objects.instance [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'flavor' on Instance uuid 87961eb2-d977-48c1-9eaf-c371eac9bbb1 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1196.538901] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.539249] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.539419] env[62183]: DEBUG nova.objects.instance [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lazy-loading 'resources' on Instance uuid 1b411fab-7728-4b29-ad9f-ce22a2face55 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1196.921688] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1196.921965] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37653064-6330-43bb-8c2d-4798e8a0d5cd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.929478] env[62183]: DEBUG oslo_vmware.api [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1196.929478] env[62183]: value = "task-1387617" [ 1196.929478] env[62183]: _type = "Task" [ 1196.929478] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.937339] env[62183]: DEBUG oslo_vmware.api [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387617, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.102892] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910c3de3-eac3-45a5-9db2-6b7666331a69 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.110354] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ffc1b1-4734-4ddf-975d-0e4b329a7ce9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.142096] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b296438-8607-47ac-956a-500ac6c37866 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.149125] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed39fbb7-8793-48eb-a2d5-17c7f5766089 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.161824] env[62183]: DEBUG nova.compute.provider_tree [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1197.439290] env[62183]: DEBUG oslo_vmware.api [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387617, 'name': PowerOffVM_Task, 'duration_secs': 0.17613} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.439574] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1197.439756] env[62183]: DEBUG nova.compute.manager [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1197.440503] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be9ede4-e48a-4a71-b7b1-906de8593644 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.504382] env[62183]: DEBUG nova.compute.manager [req-a06df3e0-af76-4252-99d1-260fc88ae381 req-0538e9e6-a2d3-4bec-8012-c2a441ed90ec service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received event network-vif-plugged-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1197.504616] env[62183]: DEBUG oslo_concurrency.lockutils [req-a06df3e0-af76-4252-99d1-260fc88ae381 req-0538e9e6-a2d3-4bec-8012-c2a441ed90ec service nova] Acquiring lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.504872] env[62183]: DEBUG oslo_concurrency.lockutils [req-a06df3e0-af76-4252-99d1-260fc88ae381 req-0538e9e6-a2d3-4bec-8012-c2a441ed90ec service nova] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.505134] env[62183]: DEBUG oslo_concurrency.lockutils [req-a06df3e0-af76-4252-99d1-260fc88ae381 req-0538e9e6-a2d3-4bec-8012-c2a441ed90ec service nova] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.505332] env[62183]: DEBUG nova.compute.manager [req-a06df3e0-af76-4252-99d1-260fc88ae381 req-0538e9e6-a2d3-4bec-8012-c2a441ed90ec service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] No waiting events found dispatching network-vif-plugged-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1197.505491] env[62183]: WARNING nova.compute.manager [req-a06df3e0-af76-4252-99d1-260fc88ae381 req-0538e9e6-a2d3-4bec-8012-c2a441ed90ec service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received unexpected event network-vif-plugged-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a for instance with vm_state shelved_offloaded and task_state spawning. [ 1197.589542] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.589817] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.589926] env[62183]: DEBUG nova.network.neutron [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1197.664530] env[62183]: DEBUG nova.scheduler.client.report [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1197.950845] env[62183]: DEBUG oslo_concurrency.lockutils [None req-e1fc3707-87fe-47a4-8933-b3ceb7b0cfd0 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.543s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.169504] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.630s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.192883] env[62183]: INFO nova.scheduler.client.report [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Deleted allocations for instance 1b411fab-7728-4b29-ad9f-ce22a2face55 [ 1198.345310] env[62183]: DEBUG nova.network.neutron [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating instance_info_cache with network_info: [{"id": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "address": "fa:16:3e:80:88:09", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ebdb1ac-70", "ovs_interfaceid": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.701751] env[62183]: DEBUG oslo_concurrency.lockutils [None req-5ba356b7-e503-4973-be8f-570ff37d88d4 tempest-AttachVolumeTestJSON-2067454659 tempest-AttachVolumeTestJSON-2067454659-project-member] Lock "1b411fab-7728-4b29-ad9f-ce22a2face55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.381s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.848236] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.874504] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f1a60520f7553cb3e764f810dde936de',container_format='bare',created_at=2024-10-31T11:16:37Z,direct_url=,disk_format='vmdk',id=72a931cc-90a1-49b8-8a97-5970cd9b0d17,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-36134411-shelved',owner='88022032e6e04a4f96bc49c2ca5ede29',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-31T11:16:50Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1198.874910] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1198.875155] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1198.875358] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1198.875511] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1198.875660] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1198.875879] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1198.876053] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1198.876227] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1198.876404] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1198.876583] env[62183]: DEBUG nova.virt.hardware [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1198.877461] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a45ca2-9f50-4080-b29d-7dade3361442 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.885377] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcb7074-c508-4692-94df-bb1875a7d496 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.898382] env[62183]: DEBUG nova.objects.instance [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'flavor' on Instance uuid 87961eb2-d977-48c1-9eaf-c371eac9bbb1 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1198.899837] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:88:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ff90ec9-3c7e-4e76-b409-fcf37fc588d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ebdb1ac-70ed-4649-95b0-4d71bd7be73a', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1198.907067] env[62183]: DEBUG oslo.service.loopingcall [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.907505] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1198.907711] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-250a8a25-0697-46cd-9a9b-76e3862adc0c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.927207] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1198.927207] env[62183]: value = "task-1387619" [ 1198.927207] env[62183]: _type = "Task" [ 1198.927207] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.934465] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387619, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.410855] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.411060] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.411255] env[62183]: DEBUG nova.network.neutron [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1199.411441] env[62183]: DEBUG nova.objects.instance [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'info_cache' on Instance uuid 87961eb2-d977-48c1-9eaf-c371eac9bbb1 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.436529] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387619, 'name': CreateVM_Task, 'duration_secs': 0.350152} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.436686] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1199.437312] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.437481] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "[datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.437880] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1199.438146] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83511591-1473-4a75-a41f-878f06543c33 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.442582] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1199.442582] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52f71047-ac8c-7411-0e31-08f3dc33e045" [ 1199.442582] env[62183]: _type = "Task" [ 1199.442582] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.449800] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52f71047-ac8c-7411-0e31-08f3dc33e045, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.534113] env[62183]: DEBUG nova.compute.manager [req-3d7899de-6736-4bec-a439-97d1ba359c22 req-b6f27d24-c1d8-429e-87d6-d5fb4b1d58df service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received event network-changed-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.534363] env[62183]: DEBUG nova.compute.manager [req-3d7899de-6736-4bec-a439-97d1ba359c22 req-b6f27d24-c1d8-429e-87d6-d5fb4b1d58df service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Refreshing instance network info cache due to event network-changed-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1199.534543] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d7899de-6736-4bec-a439-97d1ba359c22 req-b6f27d24-c1d8-429e-87d6-d5fb4b1d58df service nova] Acquiring lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.534690] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d7899de-6736-4bec-a439-97d1ba359c22 req-b6f27d24-c1d8-429e-87d6-d5fb4b1d58df service nova] Acquired lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.534941] env[62183]: DEBUG nova.network.neutron [req-3d7899de-6736-4bec-a439-97d1ba359c22 req-b6f27d24-c1d8-429e-87d6-d5fb4b1d58df service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Refreshing network info cache for port 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1199.915109] env[62183]: DEBUG nova.objects.base [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Object Instance<87961eb2-d977-48c1-9eaf-c371eac9bbb1> lazy-loaded attributes: flavor,info_cache {{(pid=62183) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1199.952182] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "[datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.952441] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Processing image 72a931cc-90a1-49b8-8a97-5970cd9b0d17 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1199.952677] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17/72a931cc-90a1-49b8-8a97-5970cd9b0d17.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.952824] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquired lock "[datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17/72a931cc-90a1-49b8-8a97-5970cd9b0d17.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.953019] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1199.953268] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-515b5c68-bc2a-4abc-89e5-8dd3403c3954 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.969753] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1199.969943] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1199.970622] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9115eddd-a031-4d65-a675-e16f0b117173 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.975478] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1199.975478] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]5213728b-3b1d-13e9-ec9f-90e54f1aee43" [ 1199.975478] env[62183]: _type = "Task" [ 1199.975478] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.983039] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]5213728b-3b1d-13e9-ec9f-90e54f1aee43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.287360] env[62183]: DEBUG nova.network.neutron [req-3d7899de-6736-4bec-a439-97d1ba359c22 req-b6f27d24-c1d8-429e-87d6-d5fb4b1d58df service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updated VIF entry in instance network info cache for port 4ebdb1ac-70ed-4649-95b0-4d71bd7be73a. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1200.287727] env[62183]: DEBUG nova.network.neutron [req-3d7899de-6736-4bec-a439-97d1ba359c22 req-b6f27d24-c1d8-429e-87d6-d5fb4b1d58df service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating instance_info_cache with network_info: [{"id": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "address": "fa:16:3e:80:88:09", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ebdb1ac-70", "ovs_interfaceid": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.485074] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Preparing fetch location {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1200.485333] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Fetch image to [datastore1] OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9/OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9.vmdk {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1200.485516] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Downloading stream optimized image 72a931cc-90a1-49b8-8a97-5970cd9b0d17 to [datastore1] OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9/OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9.vmdk on the data store datastore1 as vApp {{(pid=62183) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1200.485689] env[62183]: DEBUG nova.virt.vmwareapi.images [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Downloading image file data 72a931cc-90a1-49b8-8a97-5970cd9b0d17 to the ESX as VM named 'OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9' {{(pid=62183) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1200.557913] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1200.557913] env[62183]: value = "resgroup-9" [ 1200.557913] env[62183]: _type = "ResourcePool" [ 1200.557913] env[62183]: }. {{(pid=62183) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1200.558327] env[62183]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-1da56a05-ef26-4101-a112-01fde25aaafa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.585045] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease: (returnval){ [ 1200.585045] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525903f8-f5cc-9e0a-1630-517955fd4b56" [ 1200.585045] env[62183]: _type = "HttpNfcLease" [ 1200.585045] env[62183]: } obtained for vApp import into resource pool (val){ [ 1200.585045] env[62183]: value = "resgroup-9" [ 1200.585045] env[62183]: _type = "ResourcePool" [ 1200.585045] env[62183]: }. {{(pid=62183) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1200.585315] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the lease: (returnval){ [ 1200.585315] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525903f8-f5cc-9e0a-1630-517955fd4b56" [ 1200.585315] env[62183]: _type = "HttpNfcLease" [ 1200.585315] env[62183]: } to be ready. {{(pid=62183) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1200.591358] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1200.591358] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525903f8-f5cc-9e0a-1630-517955fd4b56" [ 1200.591358] env[62183]: _type = "HttpNfcLease" [ 1200.591358] env[62183]: } is initializing. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1200.667103] env[62183]: DEBUG nova.network.neutron [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Updating instance_info_cache with network_info: [{"id": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "address": "fa:16:3e:fa:b8:09", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae7b0413-aa", "ovs_interfaceid": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.790701] env[62183]: DEBUG oslo_concurrency.lockutils [req-3d7899de-6736-4bec-a439-97d1ba359c22 req-b6f27d24-c1d8-429e-87d6-d5fb4b1d58df service nova] Releasing lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.094593] env[62183]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1201.094593] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525903f8-f5cc-9e0a-1630-517955fd4b56" [ 1201.094593] env[62183]: _type = "HttpNfcLease" [ 1201.094593] env[62183]: } is ready. {{(pid=62183) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1201.095080] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1201.095080] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525903f8-f5cc-9e0a-1630-517955fd4b56" [ 1201.095080] env[62183]: _type = "HttpNfcLease" [ 1201.095080] env[62183]: }. {{(pid=62183) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1201.095592] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f60096-e39b-4ba8-a8b8-501b7889f46d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.102615] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204c05b-297f-4b8e-8c0b-42c44fcdedea/disk-0.vmdk from lease info. {{(pid=62183) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1201.102795] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204c05b-297f-4b8e-8c0b-42c44fcdedea/disk-0.vmdk. {{(pid=62183) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1201.165077] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0da0edef-ffd3-42fc-8363-8079551e6c88 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.169713] env[62183]: DEBUG oslo_concurrency.lockutils [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.673612] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1201.674023] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-605a19a9-b78f-40a3-b96b-72d96195b957 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.681026] env[62183]: DEBUG oslo_vmware.api [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1201.681026] env[62183]: value = "task-1387621" [ 1201.681026] env[62183]: _type = "Task" [ 1201.681026] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.688666] env[62183]: DEBUG oslo_vmware.api [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387621, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.190212] env[62183]: DEBUG oslo_vmware.api [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387621, 'name': PowerOnVM_Task, 'duration_secs': 0.44079} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.191759] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1202.191971] env[62183]: DEBUG nova.compute.manager [None req-6c643173-5c92-497f-b52c-91e0cb3358ff tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1202.192773] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850f226d-4a23-4e11-8bd4-918838b9d66a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.281722] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Completed reading data from the image iterator. {{(pid=62183) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1202.281952] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204c05b-297f-4b8e-8c0b-42c44fcdedea/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1202.282888] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c79309-98b8-46b9-af08-bc9e9a705321 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.289631] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204c05b-297f-4b8e-8c0b-42c44fcdedea/disk-0.vmdk is in state: ready. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1202.289793] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204c05b-297f-4b8e-8c0b-42c44fcdedea/disk-0.vmdk. {{(pid=62183) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1202.290035] env[62183]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-1957be74-2d6a-4135-a27c-f94ec7aae923 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.495024] env[62183]: DEBUG oslo_vmware.rw_handles [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204c05b-297f-4b8e-8c0b-42c44fcdedea/disk-0.vmdk. {{(pid=62183) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1202.495301] env[62183]: INFO nova.virt.vmwareapi.images [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Downloaded image file data 72a931cc-90a1-49b8-8a97-5970cd9b0d17 [ 1202.496155] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ec215d-5dd0-4ee0-bc1e-7c0ed174b953 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.511923] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4201a13c-f27f-4d26-8a25-75d1b950d320 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.543981] env[62183]: INFO nova.virt.vmwareapi.images [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] The imported VM was unregistered [ 1202.546524] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Caching image {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1202.546766] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Creating directory with path [datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17 {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1202.547057] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-982b85f6-eb4e-4a34-96aa-e6362e4bcf5a {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.557396] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Created directory with path [datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17 {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1202.557579] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9/OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9.vmdk to [datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17/72a931cc-90a1-49b8-8a97-5970cd9b0d17.vmdk. {{(pid=62183) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1202.557815] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-6eb6f6d8-a362-4881-b3f2-07879a5fb800 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.563741] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1202.563741] env[62183]: value = "task-1387624" [ 1202.563741] env[62183]: _type = "Task" [ 1202.563741] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.571604] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387624, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.073554] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387624, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.574254] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387624, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.976545] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88fc4ca-8e44-4a46-b2c7-420744aad3dd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.984521] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3daca381-c54b-4eb1-9e53-110d311ce01f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Suspending the VM {{(pid=62183) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1203.984889] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-a11cf12e-eb5e-43f5-a43f-2404989aefaa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.993184] env[62183]: DEBUG oslo_vmware.api [None req-3daca381-c54b-4eb1-9e53-110d311ce01f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1203.993184] env[62183]: value = "task-1387625" [ 1203.993184] env[62183]: _type = "Task" [ 1203.993184] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.002913] env[62183]: DEBUG oslo_vmware.api [None req-3daca381-c54b-4eb1-9e53-110d311ce01f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387625, 'name': SuspendVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.076153] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387624, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.504705] env[62183]: DEBUG oslo_vmware.api [None req-3daca381-c54b-4eb1-9e53-110d311ce01f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387625, 'name': SuspendVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.574682] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387624, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.003446] env[62183]: DEBUG oslo_vmware.api [None req-3daca381-c54b-4eb1-9e53-110d311ce01f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387625, 'name': SuspendVM_Task, 'duration_secs': 0.645247} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.003729] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-3daca381-c54b-4eb1-9e53-110d311ce01f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Suspended the VM {{(pid=62183) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1205.003913] env[62183]: DEBUG nova.compute.manager [None req-3daca381-c54b-4eb1-9e53-110d311ce01f tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1205.004697] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc13f0cf-0d9c-4270-9b2d-113b8aadefdd {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.075792] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387624, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.273489} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.076084] env[62183]: INFO nova.virt.vmwareapi.ds_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9/OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9.vmdk to [datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17/72a931cc-90a1-49b8-8a97-5970cd9b0d17.vmdk. [ 1205.076286] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Cleaning up location [datastore1] OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1205.076451] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_f4f26241-d016-4bab-af79-916ddf7836a9 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1205.076700] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae31644c-3278-4ae3-8679-c06bfb726fe9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.083125] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1205.083125] env[62183]: value = "task-1387627" [ 1205.083125] env[62183]: _type = "Task" [ 1205.083125] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.089849] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387627, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.593246] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387627, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038418} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.593601] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1205.593601] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Releasing lock "[datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17/72a931cc-90a1-49b8-8a97-5970cd9b0d17.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1205.593840] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17/72a931cc-90a1-49b8-8a97-5970cd9b0d17.vmdk to [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9/2d6f015c-5189-4a0b-bb79-3180d557e8f9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1205.594115] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c619a8d-c1fe-4d73-9394-17f5dce28b25 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.602280] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1205.602280] env[62183]: value = "task-1387628" [ 1205.602280] env[62183]: _type = "Task" [ 1205.602280] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.609704] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387628, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.112842] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387628, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.337439] env[62183]: INFO nova.compute.manager [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Resuming [ 1206.338191] env[62183]: DEBUG nova.objects.instance [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'flavor' on Instance uuid 87961eb2-d977-48c1-9eaf-c371eac9bbb1 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1206.613602] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387628, 'name': CopyVirtualDisk_Task} progress is 29%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.867682] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1206.867861] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1207.114132] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387628, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.347270] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.347495] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquired lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.347764] env[62183]: DEBUG nova.network.neutron [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1207.372399] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1207.372582] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Starting heal instance info cache {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1207.372682] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Rebuilding the list of instances to heal {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1207.614693] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387628, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.877090] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.877090] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquired lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.877090] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Forcefully refreshing network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1207.877090] env[62183]: DEBUG nova.objects.instance [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lazy-loading 'info_cache' on Instance uuid 2d6f015c-5189-4a0b-bb79-3180d557e8f9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.086294] env[62183]: DEBUG nova.network.neutron [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Updating instance_info_cache with network_info: [{"id": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "address": "fa:16:3e:fa:b8:09", "network": {"id": "431f297c-735a-4378-af10-799e808d928b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1621154820-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "34581a9b5c1943eabc13bbb300a0f086", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8800a981-a89e-42e4-8be9-cace419ba9cb", "external-id": "nsx-vlan-transportzone-962", "segmentation_id": 962, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae7b0413-aa", "ovs_interfaceid": "ae7b0413-aa72-4f0c-ab0b-f5c62d874948", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1208.115914] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387628, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.588952] env[62183]: DEBUG oslo_concurrency.lockutils [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Releasing lock "refresh_cache-87961eb2-d977-48c1-9eaf-c371eac9bbb1" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.589915] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce564f8e-da94-4476-8993-0564fa1193f5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.596698] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Resuming the VM {{(pid=62183) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1208.596954] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b76ece3a-3d0e-4cfb-930b-0aca8d676699 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.602731] env[62183]: DEBUG oslo_vmware.api [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1208.602731] env[62183]: value = "task-1387629" [ 1208.602731] env[62183]: _type = "Task" [ 1208.602731] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.613152] env[62183]: DEBUG oslo_vmware.api [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.616393] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387628, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.553795} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.616671] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/72a931cc-90a1-49b8-8a97-5970cd9b0d17/72a931cc-90a1-49b8-8a97-5970cd9b0d17.vmdk to [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9/2d6f015c-5189-4a0b-bb79-3180d557e8f9.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1208.617405] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2260674-0eec-4b15-a8d8-04edd535bae6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.639749] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9/2d6f015c-5189-4a0b-bb79-3180d557e8f9.vmdk or device None with type streamOptimized {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1208.640061] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74ce73a0-4ebb-4873-9a98-c919a23891c7 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.659148] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1208.659148] env[62183]: value = "task-1387630" [ 1208.659148] env[62183]: _type = "Task" [ 1208.659148] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.667271] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387630, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.118568] env[62183]: DEBUG oslo_vmware.api [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387629, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.170940] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387630, 'name': ReconfigVM_Task, 'duration_secs': 0.420892} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.171219] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9/2d6f015c-5189-4a0b-bb79-3180d557e8f9.vmdk or device None with type streamOptimized {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1209.171948] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e254482b-a37b-4dfe-bd4b-2d6945511fdf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.181650] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1209.181650] env[62183]: value = "task-1387631" [ 1209.181650] env[62183]: _type = "Task" [ 1209.181650] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.194141] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387631, 'name': Rename_Task} progress is 6%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.608799] env[62183]: DEBUG nova.network.neutron [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating instance_info_cache with network_info: [{"id": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "address": "fa:16:3e:80:88:09", "network": {"id": "38ebfd1a-5d89-4757-955c-e4000de54baf", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1821594759-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88022032e6e04a4f96bc49c2ca5ede29", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ff90ec9-3c7e-4e76-b409-fcf37fc588d8", "external-id": "nsx-vlan-transportzone-475", "segmentation_id": 475, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ebdb1ac-70", "ovs_interfaceid": "4ebdb1ac-70ed-4649-95b0-4d71bd7be73a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.615874] env[62183]: DEBUG oslo_vmware.api [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387629, 'name': PowerOnVM_Task, 'duration_secs': 0.538564} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.616134] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Resumed the VM {{(pid=62183) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1209.616322] env[62183]: DEBUG nova.compute.manager [None req-95c4b8ee-7545-40dc-a31a-3f82cd660481 tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1209.617077] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2e052a-0d05-4d7c-9461-26c18a866806 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.691011] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387631, 'name': Rename_Task, 'duration_secs': 0.218368} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.691354] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1209.691630] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c416e330-f85a-4b32-a636-be978b4d778c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.697599] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1209.697599] env[62183]: value = "task-1387632" [ 1209.697599] env[62183]: _type = "Task" [ 1209.697599] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.704809] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387632, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.112056] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Releasing lock "refresh_cache-2d6f015c-5189-4a0b-bb79-3180d557e8f9" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1210.112214] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updated the network info_cache for instance {{(pid=62183) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1210.112358] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.112512] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.112660] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.112806] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.112943] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.113104] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.113236] env[62183]: DEBUG nova.compute.manager [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62183) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1210.113381] env[62183]: DEBUG oslo_service.periodic_task [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62183) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.207814] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387632, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.616465] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.616705] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.616705] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.616893] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62183) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1210.617938] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83cc3ee-95d9-418a-af3c-1bd0dabac2ae {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.626958] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3336baa-61aa-471f-8cc3-315c526c7ddc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.641364] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c058299-2500-44ba-a39c-1f59fd66c6df {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.648088] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871b34bb-7668-4503-bfdf-adf522c694b8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.678283] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180961MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62183) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1210.678536] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.678624] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.707599] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387632, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.111466] env[62183]: DEBUG oslo_concurrency.lockutils [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.111790] env[62183]: DEBUG oslo_concurrency.lockutils [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.111979] env[62183]: DEBUG oslo_concurrency.lockutils [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.112193] env[62183]: DEBUG oslo_concurrency.lockutils [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.112382] env[62183]: DEBUG oslo_concurrency.lockutils [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.114649] env[62183]: INFO nova.compute.manager [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Terminating instance [ 1211.116525] env[62183]: DEBUG nova.compute.manager [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1211.116730] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1211.117559] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a52ecc5-7532-4b53-bfaf-b78927605f08 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.125179] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1211.125430] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb6bd52b-aae0-45db-be3f-eb56d6e92ec5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.131190] env[62183]: DEBUG oslo_vmware.api [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1211.131190] env[62183]: value = "task-1387633" [ 1211.131190] env[62183]: _type = "Task" [ 1211.131190] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.138684] env[62183]: DEBUG oslo_vmware.api [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.209034] env[62183]: DEBUG oslo_vmware.api [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387632, 'name': PowerOnVM_Task, 'duration_secs': 1.365696} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.209427] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1211.309539] env[62183]: DEBUG nova.compute.manager [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1211.310534] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14eef979-fd07-4c8f-814a-a2672de76332 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.640781] env[62183]: DEBUG oslo_vmware.api [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387633, 'name': PowerOffVM_Task, 'duration_secs': 0.235648} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.641015] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1211.641193] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1211.641441] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aeda18ae-eba2-426c-b78d-a5878f875cbe {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.702886] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 87961eb2-d977-48c1-9eaf-c371eac9bbb1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1211.703410] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Instance 2d6f015c-5189-4a0b-bb79-3180d557e8f9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62183) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1211.703410] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1211.703410] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62183) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1211.739326] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709b6c39-aa5c-4c0b-a01d-1dc3729d5881 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.747537] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a2312e-3043-4fab-be98-bbbd2d3943f2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.779546] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7cc63cd-2bb9-425f-9646-97dc2ce32f84 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.782183] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1211.782384] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1211.782566] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleting the datastore file [datastore2] 87961eb2-d977-48c1-9eaf-c371eac9bbb1 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1211.782807] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-efa07171-7b6b-46d8-b490-87771cf37986 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.790420] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54274af8-779c-422f-bcb0-b34b1a7c9b68 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.794296] env[62183]: DEBUG oslo_vmware.api [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for the task: (returnval){ [ 1211.794296] env[62183]: value = "task-1387635" [ 1211.794296] env[62183]: _type = "Task" [ 1211.794296] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.805694] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1211.812548] env[62183]: DEBUG oslo_vmware.api [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387635, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.826102] env[62183]: DEBUG oslo_concurrency.lockutils [None req-ec96238b-0357-4c3e-b51e-7a34b2cefdac tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.610s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.303884] env[62183]: DEBUG oslo_vmware.api [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Task: {'id': task-1387635, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14933} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.304216] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1212.304427] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1212.304600] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1212.304778] env[62183]: INFO nova.compute.manager [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1212.305061] env[62183]: DEBUG oslo.service.loopingcall [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1212.305268] env[62183]: DEBUG nova.compute.manager [-] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1212.305361] env[62183]: DEBUG nova.network.neutron [-] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1212.337465] env[62183]: DEBUG nova.scheduler.client.report [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 150 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1212.337794] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 150 to 151 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1212.338062] env[62183]: DEBUG nova.compute.provider_tree [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1212.408968] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.409286] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.409508] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.409695] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.409867] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.411892] env[62183]: INFO nova.compute.manager [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Terminating instance [ 1212.413622] env[62183]: DEBUG nova.compute.manager [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1212.413813] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1212.414665] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ca1e77-75f6-4907-be56-d0b8c0c23734 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.422569] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1212.422799] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6b3d45a-915c-4d4d-9db4-98e56d35fd8e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.428588] env[62183]: DEBUG oslo_vmware.api [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1212.428588] env[62183]: value = "task-1387636" [ 1212.428588] env[62183]: _type = "Task" [ 1212.428588] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.435731] env[62183]: DEBUG oslo_vmware.api [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387636, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.739522] env[62183]: DEBUG nova.compute.manager [req-913707af-2dd0-4361-89f0-1eb2209dc944 req-67294329-aacf-4170-8ff9-7f0ac7f463b1 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Received event network-vif-deleted-ae7b0413-aa72-4f0c-ab0b-f5c62d874948 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1212.739755] env[62183]: INFO nova.compute.manager [req-913707af-2dd0-4361-89f0-1eb2209dc944 req-67294329-aacf-4170-8ff9-7f0ac7f463b1 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Neutron deleted interface ae7b0413-aa72-4f0c-ab0b-f5c62d874948; detaching it from the instance and deleting it from the info cache [ 1212.739909] env[62183]: DEBUG nova.network.neutron [req-913707af-2dd0-4361-89f0-1eb2209dc944 req-67294329-aacf-4170-8ff9-7f0ac7f463b1 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.846064] env[62183]: DEBUG nova.compute.resource_tracker [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62183) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1212.846064] env[62183]: DEBUG oslo_concurrency.lockutils [None req-c153997d-a410-48b8-9ab1-75944c4287c8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.166s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.943057] env[62183]: DEBUG oslo_vmware.api [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387636, 'name': PowerOffVM_Task, 'duration_secs': 0.187477} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.943218] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1212.943373] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1212.943635] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10071dfc-45ae-41f8-9e40-391b41c63004 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.003597] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1213.003926] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Deleting contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1213.004214] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleting the datastore file [datastore1] 2d6f015c-5189-4a0b-bb79-3180d557e8f9 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1213.004558] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5a2c30b-e137-4c27-871c-556a2b5a770c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.011497] env[62183]: DEBUG oslo_vmware.api [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for the task: (returnval){ [ 1213.011497] env[62183]: value = "task-1387638" [ 1213.011497] env[62183]: _type = "Task" [ 1213.011497] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.021983] env[62183]: DEBUG oslo_vmware.api [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387638, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.219845] env[62183]: DEBUG nova.network.neutron [-] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.242452] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-486bb4a2-7b7a-485a-af88-6cd774c7d40f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.251718] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c42628-3c07-45e4-a409-54c4c63f5d99 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.274344] env[62183]: DEBUG nova.compute.manager [req-913707af-2dd0-4361-89f0-1eb2209dc944 req-67294329-aacf-4170-8ff9-7f0ac7f463b1 service nova] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Detach interface failed, port_id=ae7b0413-aa72-4f0c-ab0b-f5c62d874948, reason: Instance 87961eb2-d977-48c1-9eaf-c371eac9bbb1 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1213.521709] env[62183]: DEBUG oslo_vmware.api [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Task: {'id': task-1387638, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1458} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.521979] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1213.522471] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Deleted contents of the VM from datastore datastore1 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1213.522708] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1213.522897] env[62183]: INFO nova.compute.manager [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1213.523164] env[62183]: DEBUG oslo.service.loopingcall [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1213.523364] env[62183]: DEBUG nova.compute.manager [-] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1213.523460] env[62183]: DEBUG nova.network.neutron [-] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1213.723924] env[62183]: INFO nova.compute.manager [-] [instance: 87961eb2-d977-48c1-9eaf-c371eac9bbb1] Took 1.42 seconds to deallocate network for instance. [ 1214.230169] env[62183]: DEBUG oslo_concurrency.lockutils [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.230537] env[62183]: DEBUG oslo_concurrency.lockutils [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.230707] env[62183]: DEBUG nova.objects.instance [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lazy-loading 'resources' on Instance uuid 87961eb2-d977-48c1-9eaf-c371eac9bbb1 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.418735] env[62183]: DEBUG nova.network.neutron [-] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.765406] env[62183]: DEBUG nova.compute.manager [req-cf9c431c-cfa9-4ce7-a81d-2ec6611ea298 req-23a432b2-55a7-4723-9f88-25dbe907d9a9 service nova] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Received event network-vif-deleted-4ebdb1ac-70ed-4649-95b0-4d71bd7be73a {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1214.778062] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783e6b83-0975-4e91-81ca-f767995d39aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.785559] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820a948c-14fc-46f6-96e1-d4ab8ee255d9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.815568] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a493a24a-50a6-4f25-b162-d8248d4400de {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.824815] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c8b8e1-1f37-4d86-83f7-e92567f95cc6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.837155] env[62183]: DEBUG nova.compute.provider_tree [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1214.921089] env[62183]: INFO nova.compute.manager [-] [instance: 2d6f015c-5189-4a0b-bb79-3180d557e8f9] Took 1.40 seconds to deallocate network for instance. [ 1215.339983] env[62183]: DEBUG nova.scheduler.client.report [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1215.427221] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.844997] env[62183]: DEBUG oslo_concurrency.lockutils [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.614s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.848052] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.420s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.848052] env[62183]: DEBUG nova.objects.instance [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lazy-loading 'resources' on Instance uuid 2d6f015c-5189-4a0b-bb79-3180d557e8f9 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.862566] env[62183]: INFO nova.scheduler.client.report [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Deleted allocations for instance 87961eb2-d977-48c1-9eaf-c371eac9bbb1 [ 1216.372111] env[62183]: DEBUG oslo_concurrency.lockutils [None req-97d5c364-734f-481c-b985-b5bfe41c368e tempest-ServerActionsTestJSON-2120681273 tempest-ServerActionsTestJSON-2120681273-project-member] Lock "87961eb2-d977-48c1-9eaf-c371eac9bbb1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.260s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.383572] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d37e1b-c78e-4f19-a52b-38efe76187b0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.391548] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92fb4762-a68d-49f1-bca9-689116d02112 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.076164] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9885c61a-c1a7-4d6a-a043-b5fd8ed6337e {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.083756] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96918631-6895-429a-b11f-c310d9cd84bc {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.097012] env[62183]: DEBUG nova.compute.provider_tree [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed in ProviderTree for provider: 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.602985] env[62183]: DEBUG nova.scheduler.client.report [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Inventory has not changed for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1218.105695] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.258s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.128985] env[62183]: INFO nova.scheduler.client.report [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Deleted allocations for instance 2d6f015c-5189-4a0b-bb79-3180d557e8f9 [ 1218.637266] env[62183]: DEBUG oslo_concurrency.lockutils [None req-92aea993-20d3-4c5e-9f63-7e8f63f47480 tempest-ServerActionsTestOtherB-16165082 tempest-ServerActionsTestOtherB-16165082-project-member] Lock "2d6f015c-5189-4a0b-bb79-3180d557e8f9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.228s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.244954] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquiring lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.245262] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.748181] env[62183]: DEBUG nova.compute.manager [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Starting instance... {{(pid=62183) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1226.270159] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.270427] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.271977] env[62183]: INFO nova.compute.claims [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1227.305875] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb145014-b1bb-49f6-96eb-f8d57bc11d09 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.313457] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2dc0372-22b9-4064-a826-bd9549e87fe8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.341463] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ebed7a-b0a2-4409-88fc-29b2e7f0ab16 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.347936] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0bbb98-1035-4532-b61f-7ef729aab102 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.360312] env[62183]: DEBUG nova.compute.provider_tree [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1227.879300] env[62183]: ERROR nova.scheduler.client.report [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [req-e27936d4-e437-4380-885a-6649e040a70a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 09c07e5d-2ed9-41c2-be62-db0f731d0b87. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e27936d4-e437-4380-885a-6649e040a70a"}]} [ 1227.894527] env[62183]: DEBUG nova.scheduler.client.report [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Refreshing inventories for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1227.906704] env[62183]: DEBUG nova.scheduler.client.report [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updating ProviderTree inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1227.906916] env[62183]: DEBUG nova.compute.provider_tree [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1227.917802] env[62183]: DEBUG nova.scheduler.client.report [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Refreshing aggregate associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, aggregates: None {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1227.934896] env[62183]: DEBUG nova.scheduler.client.report [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Refreshing trait associations for resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62183) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1227.958780] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769765f8-577e-47ed-94f0-7647715e72e2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.965895] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9048764-3347-4a00-a05c-2dcaa7c56725 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.994791] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29522f72-ebb6-473a-a884-7722a14c6193 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.001250] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade36126-74d6-48d9-bd4c-2d3eb10e1413 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.013570] env[62183]: DEBUG nova.compute.provider_tree [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1228.541595] env[62183]: DEBUG nova.scheduler.client.report [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 152 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1228.541920] env[62183]: DEBUG nova.compute.provider_tree [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 152 to 153 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1228.542069] env[62183]: DEBUG nova.compute.provider_tree [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 176, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1229.051017] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.777s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.051017] env[62183]: DEBUG nova.compute.manager [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Start building networks asynchronously for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1229.554553] env[62183]: DEBUG nova.compute.utils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Using /dev/sd instead of None {{(pid=62183) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1229.556017] env[62183]: DEBUG nova.compute.manager [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Allocating IP information in the background. {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1229.556232] env[62183]: DEBUG nova.network.neutron [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] allocate_for_instance() {{(pid=62183) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1229.595191] env[62183]: DEBUG nova.policy [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a22ea6d6d5f44158a8c28a82d4f226f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38ab78d0546a4b67a69f68f09a3eb26b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62183) authorize /opt/stack/nova/nova/policy.py:201}} [ 1229.829945] env[62183]: DEBUG nova.network.neutron [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Successfully created port: 7325ed56-bd2c-42ce-b83b-3af25efd1394 {{(pid=62183) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1230.059772] env[62183]: DEBUG nova.compute.manager [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Start building block device mappings for instance. {{(pid=62183) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1231.069396] env[62183]: DEBUG nova.compute.manager [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Start spawning the instance on the hypervisor. {{(pid=62183) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1231.093850] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T11:04:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T11:03:45Z,direct_url=,disk_format='vmdk',id=b250b36d-7961-4edf-b9ae-4be16b3e5ad2,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='d27665f42cab4ede844047cab22995c7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T11:03:45Z,virtual_size=,visibility=), allow threads: False {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1231.094118] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Flavor limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1231.094283] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Image limits 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1231.094542] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Flavor pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1231.094605] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Image pref 0:0:0 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1231.094755] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62183) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1231.094958] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1231.095138] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1231.095308] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Got 1 possible topologies {{(pid=62183) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1231.095483] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1231.095643] env[62183]: DEBUG nova.virt.hardware [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62183) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1231.096518] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a768cdc7-6263-47ba-bf91-71ac050ac985 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.104493] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee283412-4390-4961-960a-d3c652c3d5aa {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.192960] env[62183]: DEBUG nova.compute.manager [req-e25960c1-593d-4b58-95c3-ea742d550c38 req-fead33ec-243b-498d-91f1-bfe9e9e87559 service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Received event network-vif-plugged-7325ed56-bd2c-42ce-b83b-3af25efd1394 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1231.193251] env[62183]: DEBUG oslo_concurrency.lockutils [req-e25960c1-593d-4b58-95c3-ea742d550c38 req-fead33ec-243b-498d-91f1-bfe9e9e87559 service nova] Acquiring lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.193426] env[62183]: DEBUG oslo_concurrency.lockutils [req-e25960c1-593d-4b58-95c3-ea742d550c38 req-fead33ec-243b-498d-91f1-bfe9e9e87559 service nova] Lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.193611] env[62183]: DEBUG oslo_concurrency.lockutils [req-e25960c1-593d-4b58-95c3-ea742d550c38 req-fead33ec-243b-498d-91f1-bfe9e9e87559 service nova] Lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1231.193746] env[62183]: DEBUG nova.compute.manager [req-e25960c1-593d-4b58-95c3-ea742d550c38 req-fead33ec-243b-498d-91f1-bfe9e9e87559 service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] No waiting events found dispatching network-vif-plugged-7325ed56-bd2c-42ce-b83b-3af25efd1394 {{(pid=62183) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1231.193913] env[62183]: WARNING nova.compute.manager [req-e25960c1-593d-4b58-95c3-ea742d550c38 req-fead33ec-243b-498d-91f1-bfe9e9e87559 service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Received unexpected event network-vif-plugged-7325ed56-bd2c-42ce-b83b-3af25efd1394 for instance with vm_state building and task_state spawning. [ 1231.275938] env[62183]: DEBUG nova.network.neutron [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Successfully updated port: 7325ed56-bd2c-42ce-b83b-3af25efd1394 {{(pid=62183) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1231.778836] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquiring lock "refresh_cache-4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1231.779020] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquired lock "refresh_cache-4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1231.779180] env[62183]: DEBUG nova.network.neutron [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Building network info cache for instance {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1232.310451] env[62183]: DEBUG nova.network.neutron [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Instance cache missing network info. {{(pid=62183) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1232.492734] env[62183]: DEBUG nova.network.neutron [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Updating instance_info_cache with network_info: [{"id": "7325ed56-bd2c-42ce-b83b-3af25efd1394", "address": "fa:16:3e:1e:23:b5", "network": {"id": "bbd10020-bb14-49c0-a83c-b2e21e603213", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-2090411767-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ab78d0546a4b67a69f68f09a3eb26b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7325ed56-bd", "ovs_interfaceid": "7325ed56-bd2c-42ce-b83b-3af25efd1394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.995542] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Releasing lock "refresh_cache-4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1232.995905] env[62183]: DEBUG nova.compute.manager [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Instance network_info: |[{"id": "7325ed56-bd2c-42ce-b83b-3af25efd1394", "address": "fa:16:3e:1e:23:b5", "network": {"id": "bbd10020-bb14-49c0-a83c-b2e21e603213", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-2090411767-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ab78d0546a4b67a69f68f09a3eb26b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7325ed56-bd", "ovs_interfaceid": "7325ed56-bd2c-42ce-b83b-3af25efd1394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62183) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1232.996428] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:23:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7325ed56-bd2c-42ce-b83b-3af25efd1394', 'vif_model': 'vmxnet3'}] {{(pid=62183) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1233.004087] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Creating folder: Project (38ab78d0546a4b67a69f68f09a3eb26b). Parent ref: group-v294392. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1233.004379] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9fc1a910-47dc-4502-a578-c1c8bfe3f670 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.015066] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Created folder: Project (38ab78d0546a4b67a69f68f09a3eb26b) in parent group-v294392. [ 1233.015258] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Creating folder: Instances. Parent ref: group-v294537. {{(pid=62183) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1233.015482] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d49d467e-3fc4-43ab-bb3d-de4b38da26e0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.023783] env[62183]: INFO nova.virt.vmwareapi.vm_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Created folder: Instances in parent group-v294537. [ 1233.024010] env[62183]: DEBUG oslo.service.loopingcall [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1233.024195] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Creating VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1233.024383] env[62183]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd635f09-c8f7-4603-9a29-0f8b1c57f132 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.041807] env[62183]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1233.041807] env[62183]: value = "task-1387642" [ 1233.041807] env[62183]: _type = "Task" [ 1233.041807] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.048573] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387642, 'name': CreateVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.218725] env[62183]: DEBUG nova.compute.manager [req-302b748d-b548-4075-9b99-e69a77e6c500 req-d4dff23f-cfa9-45a0-a415-7586bd11aa8a service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Received event network-changed-7325ed56-bd2c-42ce-b83b-3af25efd1394 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1233.218981] env[62183]: DEBUG nova.compute.manager [req-302b748d-b548-4075-9b99-e69a77e6c500 req-d4dff23f-cfa9-45a0-a415-7586bd11aa8a service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Refreshing instance network info cache due to event network-changed-7325ed56-bd2c-42ce-b83b-3af25efd1394. {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1233.219210] env[62183]: DEBUG oslo_concurrency.lockutils [req-302b748d-b548-4075-9b99-e69a77e6c500 req-d4dff23f-cfa9-45a0-a415-7586bd11aa8a service nova] Acquiring lock "refresh_cache-4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1233.219356] env[62183]: DEBUG oslo_concurrency.lockutils [req-302b748d-b548-4075-9b99-e69a77e6c500 req-d4dff23f-cfa9-45a0-a415-7586bd11aa8a service nova] Acquired lock "refresh_cache-4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1233.219519] env[62183]: DEBUG nova.network.neutron [req-302b748d-b548-4075-9b99-e69a77e6c500 req-d4dff23f-cfa9-45a0-a415-7586bd11aa8a service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Refreshing network info cache for port 7325ed56-bd2c-42ce-b83b-3af25efd1394 {{(pid=62183) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1233.551947] env[62183]: DEBUG oslo_vmware.api [-] Task: {'id': task-1387642, 'name': CreateVM_Task, 'duration_secs': 0.281269} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.552238] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Created VM on the ESX host {{(pid=62183) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1233.552790] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1233.552969] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1233.553302] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1233.553559] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-176e363e-202d-4b9c-b5dd-c416ea1fafa3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.557732] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1233.557732] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]52340846-65cd-e126-42ea-efe5b2af2f9a" [ 1233.557732] env[62183]: _type = "Task" [ 1233.557732] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.565261] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52340846-65cd-e126-42ea-efe5b2af2f9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.904162] env[62183]: DEBUG nova.network.neutron [req-302b748d-b548-4075-9b99-e69a77e6c500 req-d4dff23f-cfa9-45a0-a415-7586bd11aa8a service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Updated VIF entry in instance network info cache for port 7325ed56-bd2c-42ce-b83b-3af25efd1394. {{(pid=62183) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1233.904551] env[62183]: DEBUG nova.network.neutron [req-302b748d-b548-4075-9b99-e69a77e6c500 req-d4dff23f-cfa9-45a0-a415-7586bd11aa8a service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Updating instance_info_cache with network_info: [{"id": "7325ed56-bd2c-42ce-b83b-3af25efd1394", "address": "fa:16:3e:1e:23:b5", "network": {"id": "bbd10020-bb14-49c0-a83c-b2e21e603213", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-2090411767-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38ab78d0546a4b67a69f68f09a3eb26b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7325ed56-bd", "ovs_interfaceid": "7325ed56-bd2c-42ce-b83b-3af25efd1394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1234.067469] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]52340846-65cd-e126-42ea-efe5b2af2f9a, 'name': SearchDatastore_Task, 'duration_secs': 0.010027} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.067742] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1234.067987] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Processing image b250b36d-7961-4edf-b9ae-4be16b3e5ad2 {{(pid=62183) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1234.068290] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.068448] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.068633] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1234.068894] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec8b2788-d44b-4542-bfe7-cd16afdb3b85 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.078615] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62183) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1234.078844] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62183) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1234.079664] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0ace1e9-526a-4ca4-bd5b-5b20d63f6e76 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.084652] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1234.084652] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]529b54a7-f5c6-c5c0-6c2c-c8b5fb252add" [ 1234.084652] env[62183]: _type = "Task" [ 1234.084652] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.091763] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529b54a7-f5c6-c5c0-6c2c-c8b5fb252add, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.407152] env[62183]: DEBUG oslo_concurrency.lockutils [req-302b748d-b548-4075-9b99-e69a77e6c500 req-d4dff23f-cfa9-45a0-a415-7586bd11aa8a service nova] Releasing lock "refresh_cache-4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1234.594858] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]529b54a7-f5c6-c5c0-6c2c-c8b5fb252add, 'name': SearchDatastore_Task, 'duration_secs': 0.008066} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.595668] env[62183]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60a0ca2a-333c-43d0-acd2-0137a4920ed6 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.600822] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1234.600822] env[62183]: value = "session[52f03c31-072d-5d99-239b-45bf1984c659]525455e2-c52f-2ffe-8ad6-fa4a1d2a9b21" [ 1234.600822] env[62183]: _type = "Task" [ 1234.600822] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.608640] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525455e2-c52f-2ffe-8ad6-fa4a1d2a9b21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.111210] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': session[52f03c31-072d-5d99-239b-45bf1984c659]525455e2-c52f-2ffe-8ad6-fa4a1d2a9b21, 'name': SearchDatastore_Task, 'duration_secs': 0.008606} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.111475] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk" {{(pid=62183) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.111740] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68/4f31c88e-d048-4eb8-8f88-9fa68ebd2b68.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1235.111993] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fffe5e3c-d8c7-4b12-baaf-aaa6676365b5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.118511] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1235.118511] env[62183]: value = "task-1387643" [ 1235.118511] env[62183]: _type = "Task" [ 1235.118511] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.125416] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387643, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.628369] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387643, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44864} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.628675] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b250b36d-7961-4edf-b9ae-4be16b3e5ad2/b250b36d-7961-4edf-b9ae-4be16b3e5ad2.vmdk to [datastore2] 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68/4f31c88e-d048-4eb8-8f88-9fa68ebd2b68.vmdk {{(pid=62183) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1235.628851] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Extending root virtual disk to 1048576 {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1235.629113] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-76f369d9-29ac-4fc9-a95d-e37b1df6c9b8 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.635700] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1235.635700] env[62183]: value = "task-1387644" [ 1235.635700] env[62183]: _type = "Task" [ 1235.635700] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.642557] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387644, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.144914] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387644, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064504} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.145221] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Extended root virtual disk {{(pid=62183) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1236.145965] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862fe06b-ce8f-4b0a-b702-db6f2781962c {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.167448] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68/4f31c88e-d048-4eb8-8f88-9fa68ebd2b68.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1236.167683] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77c165c4-bd28-4e0a-bcd5-e859afab89e0 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.185855] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1236.185855] env[62183]: value = "task-1387645" [ 1236.185855] env[62183]: _type = "Task" [ 1236.185855] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.195157] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387645, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.694998] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387645, 'name': ReconfigVM_Task, 'duration_secs': 0.305683} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.695374] env[62183]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68/4f31c88e-d048-4eb8-8f88-9fa68ebd2b68.vmdk or device None with type sparse {{(pid=62183) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1236.695965] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3aea6534-5543-43ce-a356-88058111547d {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.702403] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1236.702403] env[62183]: value = "task-1387646" [ 1236.702403] env[62183]: _type = "Task" [ 1236.702403] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.709696] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387646, 'name': Rename_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.213013] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387646, 'name': Rename_Task, 'duration_secs': 0.13394} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.213013] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Powering on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1237.213013] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73112ef0-b84c-4c50-b58a-96a6d9f190f2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.218215] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1237.218215] env[62183]: value = "task-1387647" [ 1237.218215] env[62183]: _type = "Task" [ 1237.218215] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.226153] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387647, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.728481] env[62183]: DEBUG oslo_vmware.api [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387647, 'name': PowerOnVM_Task, 'duration_secs': 0.425109} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.728889] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Powered on the VM {{(pid=62183) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1237.728889] env[62183]: INFO nova.compute.manager [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Took 6.66 seconds to spawn the instance on the hypervisor. [ 1237.728889] env[62183]: DEBUG nova.compute.manager [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Checking state {{(pid=62183) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1237.729655] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf710d0-aebf-4d3d-ae2a-75f843fbe1ad {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.246679] env[62183]: INFO nova.compute.manager [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Took 11.99 seconds to build instance. [ 1238.748416] env[62183]: DEBUG oslo_concurrency.lockutils [None req-7a4f50aa-229c-48cb-baf4-0bf61afd8e54 tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.503s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.972348] env[62183]: DEBUG oslo_concurrency.lockutils [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquiring lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.972743] env[62183]: DEBUG oslo_concurrency.lockutils [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.973034] env[62183]: DEBUG oslo_concurrency.lockutils [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquiring lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.973283] env[62183]: DEBUG oslo_concurrency.lockutils [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.973498] env[62183]: DEBUG oslo_concurrency.lockutils [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.975879] env[62183]: INFO nova.compute.manager [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Terminating instance [ 1238.977664] env[62183]: DEBUG nova.compute.manager [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Start destroying the instance on the hypervisor. {{(pid=62183) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1238.977881] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Destroying instance {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1238.978754] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4dba660-cdc8-4d8d-b89b-33ad703b4d9b {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.986531] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Powering off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1238.986757] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a8bf7a7-6c35-4a9c-83c1-ea4c88ae83ec {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.992234] env[62183]: DEBUG oslo_vmware.api [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1238.992234] env[62183]: value = "task-1387648" [ 1238.992234] env[62183]: _type = "Task" [ 1238.992234] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.001402] env[62183]: DEBUG oslo_vmware.api [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387648, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.502425] env[62183]: DEBUG oslo_vmware.api [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387648, 'name': PowerOffVM_Task, 'duration_secs': 0.203895} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.502692] env[62183]: DEBUG nova.virt.vmwareapi.vm_util [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Powered off the VM {{(pid=62183) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1239.502872] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Unregistering the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1239.503141] env[62183]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4c715c0-5788-4879-81ed-ccff7bf3aec2 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.563510] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Unregistered the VM {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1239.563749] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Deleting contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1239.563924] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Deleting the datastore file [datastore2] 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68 {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1239.564224] env[62183]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8025cb0e-4d69-4ffb-abf7-b036ecf84c17 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.570935] env[62183]: DEBUG oslo_vmware.api [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for the task: (returnval){ [ 1239.570935] env[62183]: value = "task-1387650" [ 1239.570935] env[62183]: _type = "Task" [ 1239.570935] env[62183]: } to complete. {{(pid=62183) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.579046] env[62183]: DEBUG oslo_vmware.api [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387650, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.081385] env[62183]: DEBUG oslo_vmware.api [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Task: {'id': task-1387650, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152913} completed successfully. {{(pid=62183) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.082181] env[62183]: DEBUG nova.virt.vmwareapi.ds_util [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Deleted the datastore file {{(pid=62183) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1240.082181] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Deleted contents of the VM from datastore datastore2 {{(pid=62183) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1240.082181] env[62183]: DEBUG nova.virt.vmwareapi.vmops [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Instance destroyed {{(pid=62183) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1240.082389] env[62183]: INFO nova.compute.manager [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1240.082687] env[62183]: DEBUG oslo.service.loopingcall [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62183) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1240.082984] env[62183]: DEBUG nova.compute.manager [-] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Deallocating network for instance {{(pid=62183) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1240.083074] env[62183]: DEBUG nova.network.neutron [-] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] deallocate_for_instance() {{(pid=62183) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1240.317462] env[62183]: DEBUG nova.compute.manager [req-206e69ee-33e7-45bf-822a-531e27945b3d req-5d2ed0bf-7a45-4659-a498-6968b99295b9 service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Received event network-vif-deleted-7325ed56-bd2c-42ce-b83b-3af25efd1394 {{(pid=62183) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1240.317687] env[62183]: INFO nova.compute.manager [req-206e69ee-33e7-45bf-822a-531e27945b3d req-5d2ed0bf-7a45-4659-a498-6968b99295b9 service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Neutron deleted interface 7325ed56-bd2c-42ce-b83b-3af25efd1394; detaching it from the instance and deleting it from the info cache [ 1240.317864] env[62183]: DEBUG nova.network.neutron [req-206e69ee-33e7-45bf-822a-531e27945b3d req-5d2ed0bf-7a45-4659-a498-6968b99295b9 service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.791176] env[62183]: DEBUG nova.network.neutron [-] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Updating instance_info_cache with network_info: [] {{(pid=62183) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.820473] env[62183]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e40e17f-756b-4d92-8dc2-965fac196ecf {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.830838] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba85abd-138e-48e9-a70e-99d1f28d5aba {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.853034] env[62183]: DEBUG nova.compute.manager [req-206e69ee-33e7-45bf-822a-531e27945b3d req-5d2ed0bf-7a45-4659-a498-6968b99295b9 service nova] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Detach interface failed, port_id=7325ed56-bd2c-42ce-b83b-3af25efd1394, reason: Instance 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68 could not be found. {{(pid=62183) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1241.293815] env[62183]: INFO nova.compute.manager [-] [instance: 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68] Took 1.21 seconds to deallocate network for instance. [ 1241.800201] env[62183]: DEBUG oslo_concurrency.lockutils [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1241.800532] env[62183]: DEBUG oslo_concurrency.lockutils [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1241.800800] env[62183]: DEBUG nova.objects.instance [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lazy-loading 'resources' on Instance uuid 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68 {{(pid=62183) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1242.337057] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf1e875-0d04-41c8-a142-39ba4b5a660f {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.344129] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf310fb-da9f-44ae-b924-2fe4155ae6f9 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.372644] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfcdb2f-e98b-40a8-b048-85cbc71698a3 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.379443] env[62183]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a60ecc-c702-44fc-8f06-06d37a35bed5 {{(pid=62183) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.392910] env[62183]: DEBUG nova.compute.provider_tree [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1242.921677] env[62183]: DEBUG nova.scheduler.client.report [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updated inventory for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with generation 153 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1242.921956] env[62183]: DEBUG nova.compute.provider_tree [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updating resource provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 generation from 153 to 154 during operation: update_inventory {{(pid=62183) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1242.922158] env[62183]: DEBUG nova.compute.provider_tree [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Updating inventory in ProviderTree for provider 09c07e5d-2ed9-41c2-be62-db0f731d0b87 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62183) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1243.426828] env[62183]: DEBUG oslo_concurrency.lockutils [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.626s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.444269] env[62183]: INFO nova.scheduler.client.report [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Deleted allocations for instance 4f31c88e-d048-4eb8-8f88-9fa68ebd2b68 [ 1243.952860] env[62183]: DEBUG oslo_concurrency.lockutils [None req-559b6fbd-8cf9-4815-9d1c-cfc2562f15fb tempest-ServerMetadataNegativeTestJSON-1015503976 tempest-ServerMetadataNegativeTestJSON-1015503976-project-member] Lock "4f31c88e-d048-4eb8-8f88-9fa68ebd2b68" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.980s {{(pid=62183) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}